A Performance Study of Veritas Storage Foundation Cluster File System in a NFS file serving environment. March 2008

Size: px
Start display at page:

Download "A Performance Study of Veritas Storage Foundation Cluster File System in a NFS file serving environment. March 2008"

Transcription

1 A Performance Study of Veritas Storage Foundation Cluster File System in a NFS file serving environment 1

2 TABLE OF CONTENTS EXECUTIVE SUMMARY... 3 SCOPE & AUDIENCE... 4 INTRODUCTION... 5 About Veritas Storage Foundation Cluster File System (CFS)... 5 About Network File System (NFS)... 5 BENCHMARK CONFIGURATION... 6 Hardware Configuration... 6 Software Configuration... 6 BENCHMARK DETAILS... 8 Tested configurations... 8 BENCHMARK RESULTS /14

3 Executive Summary This paper describes the scalability of the Veritas Storage Foundation Cluster File System 5.0 (CFS) on Linux for a high volume transactional workload with a mix of file operations. Files were served using the Network File System (NFS) protocol for workloads based on the widely used and well understood SPECsfs benchmark (SPEC.org). In these tests, the benchmark suite from SPEC was used as a load generator to allow for measuring the scalability factors of CFS. Scalability measurements were taken for configurations ranging from 1-node up to 16-nodes. To ensure accurate results measurements were taken at different load points throughout the test. This led to two distinct problems. The first problem raised is a direct result of bottlenecking resulting from the configurations for the 1 and 2 node tests. With these smaller configurations, the arrays themselves became a bottleneck and maximum throughput to the arrays was reached long before the CPUs became saturated. With larger configurations, more arrays and disks were brought into play, which improved the storage configurations ability to handle higher throughput levels. The reason for this was that throughout the test a 1 to 1 relationship between nodes and arrays was used. This balanced configuration was used to ensure that the tests were performed on a realistic environment and to simulate a growing environment. The second problem with our tests is that measuring throughput at a sustained 100% CPU utilization is unrealistic in real life environments. As a result, we decided to adjust I/O throughput for each test until a more realistic 80% CPU utilization and base our tests on the amount of throughput required to reach an 80% utilization as the target for each configuration we tested. At this level of CPU utilization, the storage configurations used in these tests could be compared more easily. The scalability factors that we obtained by measuring the level of throughput required to cause 80% CPU utilization was almost linear across the range of configurations. The 16-node configuration scaled to 91% of its theoretical maximum based on 16 times the throughput of the 1-node configuration. With a scalability factor of 91% at 16-nodes, the Veritas Storage Foundation Cluster File System 5.0 provides the most scalable infrastructure for complex workloads of any clustered file system for the Linux operating system. 3/14

4 Scope & Audience This paper explains the value that Veritas Storage Foundation software can bring to customers currently using, or planning to use, Network File System (NFS) for data sharing in a performance critical environment. The paper discusses the results of a performance benchmark for a file serving workload. It does not explain in detail how to set up the configurations studied. Rather, it only describes the outcome and implications of the performance benchmark. The paper is intended for engineering/architecture and operations groups for storage and for UNIX servers in large enterprise data centers who are deploying or evaluating Veritas Storage Foundation Cluster File System. This paper assumes basic knowledge of NFS and Veritas Storage Foundation Cluster File System as well an understanding of data center storage management tools, such as SANs (Storage Area Networks), logical volumes, and file systems. 4/14

5 Introduction This paper gives insight into how the Veritas Storage Foundation Cluster File System (CFS) solution can provide scalability in high-performance file serving environments. In some environments Veritas Storage Foundation Cluster File System has become a natural replacement for NFS (Network File System), but as this paper shows, using NFS together with CFS can provide high availability and high performance for your file serving needs. The Veritas Storage Foundation Cluster File System solution guarantees availability by leveraging the leading cross-platform clustering solution, Veritas Cluster Server (VCS). Close integration with VCS enables the Veritas Storage Foundation Cluster File System solution to provide a highly available cluster file system infrastructure across multiple servers even when individual components or servers fail. As an NFS replacement, CFS is able to increase availability and performance by leveraging the SAN (Storage Area Network) for data transfers instead of slower TCP/IP based networks. When used in conjunction with NFS, CFS provides a highly available, scalable platform for environments with high transaction throughput. This document illustrates how well CFS scales for transactional workloads like file serving. While the benchmark configuration in this study focuses on CFS scalability for NFS file sharing between servers and clients, the Veritas Cluster File System scales equally well for parallel applications running locally. About Veritas Storage Foundation Cluster File System (CFS) Veritas Storage Foundation Cluster File System provides an integrated solution for shared file environments. The solution includes Veritas Cluster File System, Cluster Volume Manager and Cluster Server to help implement robust, manageable, and scalable shared file solutions. With VERITAS Storage Foundation Cluster File System, cluster-wide volume and file system configuration allows for simplified management; and extending clusters is simplified as new servers adopt cluster-wide configurations. Supports the POSIX standard and maintains cache-coherency across all cluster nodes. About Network File System (NFS) NFS is the most popular network protocol on the market to share data between UNIX and Linux servers; it was first introduced in 1984 by Sun Microsystems and later was made an official standard for transferring files between computers over the network. It utilizes TCP/IP communicate between clients and servers to concurrently share files over the network. It does not fully support the POSIX standard and is therefore not suitable for environments where data integrity across nodes is demanded. 5/14

6 Benchmark configuration Hardware Configuration Servers: 16x Sun v20z each with two 2.2 GHz AMD Opteron CPUs and 8 GB RAM Clients: 32x SuperMicro each with 3.0GHz Intel Pentium 4 CPUs and 1GB of RAM Network: Each server and client was equipped with a dedicated Gigabit network adapter for the private network used to benchmark data traffic. Storage: Each server was connected to 16 Sun StorEdge 3510 arrays through a dual ported 2 Gigabit Fibre Channel. Each array was equipped with a dual ported 2 Gigabit controller, 36x 15K RPM 36GB disks and 1 GB of cache. Software Configuration Operating System and Software stack This benchmark was performed on RedHat Enterprise Linux Advanced Server 4 with update 4. The Linux kernel used during the tests was version ELsmp. Veritas Storage Foundation Cluster File System version 5.0 with maintenance pack 1 was used during this benchmark. Storage Layout Each Sun 3510 array was configured to export 6 LUNs. Each LUN was created from 6 disks in a RAID-0 (stripe) to maximize performance. The stripe size was kept at the default 32k. Veritas Volume Manager The Veritas Volume Manager layout was scaled based on the number of nodes and arrays in each test run. Following best practices for deployment in environments with high transaction rates, separate metadata and data volumes were used to minimize the risk of creating a bottleneck at the disk level for journal updates. One disk group consisting of 6 LUNs (1-node test) to 96 LUNs (16-node test) was used 8 RAID-0 volumes were striped across the disk group for file system metadata; Their size calculated as 3GB*number of arrays 8 RAID-0 volumes were striped across the disk group for data storage. Their size calculated as 100GB*number of arrays 8 Volume sets were created. Each consisting of one metadata and one data volume 6/14

7 Veritas File System All tests used 8 file systems The file systems were created with a block size of 1KB and an intent log size of 256Mb The data portion of the file systems ranged from 100 GB to 1600 GB as the environment scaled from 1-node to 16-nodes. The size was calculated as 100 GB*Number of participating nodes All file systems were mounted as cluster file systems ( mount option -o cluster ) Primary nodes for the 8 cluster file systems were rotated among the server nodes in the configuration (for example, in the 4-node test, each server node was the primary mount point for two file systems). In the 12-node and 16-node tests, only eight nodes served as primary nodes (one node for each file system); the remaining nodes were secondary on all eight file systems. NFS tuning parameters To ensure the best possible NFS performance throughout the tests the number of NFS daemons was increased to 96. Network tuning parameters The Linux networking stack is conservative when it assigns memory to resources to network buffers. For workloads generating a lot of network traffic, increasing the defaults can yield significant performance improvements. The table below lists the modified network parameters. Table 1: Network tuning parameters Tunable Value Description /proc/sys/net/core/rmem_default /proc/sys/net/core/rmem_max /proc/sys/net/core/wmem_default /proc/sys/net/core/wmem_max Default and Maximum Socket Receive Buffer size Default and Maximum Socket Send Buffer size /proc/sys/net/ipv4/tcp_rmem /proc/sys/net/ipv4/tcp_wmem Minimum, Initial and Maximum size of TCP socket recieve buffer Minimum, Initial and Maximum size of TCP socket send buffer /proc/sys/net/ipv4/tcp_sack 0 Disables selective TCP Acknowledgement /proc/sys/net/ipv4/tcp_timestamps 0 Disable TCP timestamps 7/14

8 Benchmark details The benchmark suite used to perform these tests was SPECsfs (SFS97_R1 V3.0, formerly known as LADDIS) from (Standard Performance Evaluation Company). This is a benchmark suite specifically designed to measure performance of file serving over the NFS protocol and it is the industry standard used to measure performance among NAS (Network Attached Storage) vendors. As a benchmark tool SPECsfs is widely accepted as a good load generator for a mix of small, transactional read and writes. For file systems, this presents a difficult workload to achieve good performance results, as it requires lots of metadata and data updates across a large data set. Note that while the SPECsfs benchmark was run according to the rules and regulations from SPEC this benchmark has not been submitted as an official result. There was no intention to publish official results of this benchmark as performance of this hardware configuration is not particularly interesting. The SPECsfs benchmark suite was used only as a load generator to ensure that that the file system was put under stress from a known workload while the scalability of the file system was measured. Tested configurations To demonstrate the scalability of the Veritas Storage Foundation Cluster File System product, tests were conducted in multiple configurations. The test matrix below lists all permutations of Servers, Arrays and clients used during the tests: Table 2 Test matrix Server(s) Array(s) Clients This setup allowed for linear scaling of the workload and resources available to CFS In each case, 8 file systems were created and allocated across all of the arrays and accessed by all of the clients and server nodes. To be able to compare the runs on different numbers of nodes at comparable load points, the load generation on the clients was incremented by approximately 1000 operations per second per node. Each run consisted of 15 load points. 8/14

9 Benchmark results As with all benchmarks, one should not look just at the maximum throughput, but also consider other aspects of the configuration. For example CPU and memory utilization at a given load point. The benefit of taking system resource utilization into account is that at similar loads it becomes possible to compare test results from multiple configurations. Figure 1 shows the performance of the six tested configurations, plotting Response Time on the Y-axis and Operations per Second on the X-axis. To illustrate that the performance was about the same for each configuration, the X-axis is the average of Operations per Second per server node (total operations per second divided by the number of server nodes in the test) so that all six configurations can be shown in a single graph. Figure 1: Throughput per Node vs Response Time 9/14

10 Figures Figure 2 through Figure 4 illustrate different ways of calculating CFS scalability as the configuration grew from a 1-node to 16-nodes. In Figure 2, the maximum throughput for each configuration (bar graph) and the average CPU utilization (over all server nodes) at the maximum point (line graph) is plotted. The CPU utilization line graph show that the average CPU utilization at the maximum throughput point increased with more nodes, and that the configuration with the lowest average CPU utilization was the 1-node case. Figure 2: Maximum Throughput and CPU Utilization (at Maximum Throughput) for all Configurations The main reason the cases with fewer server nodes had the lowest CPU utilization was an I/O bottleneck that was primarily a problem with one to four server nodes tests. As previously mentioned a balanced system (The number of arrays was kept equal to the number of server nodes) was maintained throughout the tests. This means that as we grew the configuration more spindles were added to the array and as a result the throughput bottleneck shifted from the array to the CPUs. 10/14

11 Tables Table 3 and Table 4 show values from a subset of the data collected where the same load level was applied for each of the configurations (about OPS/second per node). The number of average I/Os per operation and KBs moved per operation is almost identical for 1- node to 16-nodes; thus, the load on each array should be about the same. However, the iostat value for average wait (time to do an I/O including wait time) is more than four times larger in the 1-node configuration compared to the 16-node configuration. A number of tests were performed to investigate the cause of this I/O bottleneck, but no firm conclusions were reached. Table 3: Comparison of I/O performance at approximately the same throughput (about Ops/s) per Node Number Of Nodes Measured Ops per Second Measured Ops per Second per Node Response Time (ms) (R+W) I/Os per OP (R+W) KBs per OP iostat average wait time iostat average wait time Table 4: Comparison of CPU load at approximately the same throughput (about Ops/s) per Node Number Of Nodes Measured Ops per Second Measured Ops per Second per Node Response Time (ms) sar average user CPU busy sar average sys CPU busy sar average CPU busy Approx. CPU usec busy per OP /14

12 The low CPU utilization caused by I/O throughput bottlenecks in the 1 to 4 node configurations distorts the apparent scalability of the other configurations. Using maximum throughput from each configuration the scalability from 1-node to 8-nodes would be 8.12 (118030/14539). This value (8.12) is unreasonable unless it is noted that the CPU in the 1-node case was only 80.5% busy, but in the 8-node was 96.8% busy. Because the 1-node configuration could only achieve about 80% CPU utilization and also most customers would not be running systems at an average close to 100% utilization, a better assessment of scalability is to compare the different configurations at about the same (high) CPU utilization. Figure 3 is like Figure 2 except the throughput and CPU utilization are from the measured load that came closest to 80% busy for each configuration. Figure 3: Throughput and CPU Utilization at 80% CPU Busy 12/14

13 Figure 4 presents the same data as Figure 3, but instead of reporting throughput, the graph reports the scaled throughput multiple for each configuration relative to the 1-node configuration (Throughput with N-nodes divided by throughput with 1-node at about 80% busy). As can be seen in these two figures the collected data points are in all cases very close to an average CPU utilization of 80% (The line graph is almost parallel to the X-axis). In Figure 3 the measured scalability is more reasonable for comparison and shows a very high level of scalability. For example, the 1-node to 8-nodes throughput ratio is 7.37 and the 1-node to 16- node ratio is One of the reasons the values are not 8.0 and 16.0 is that the CPU busy time per operation is higher when there are more server nodes. In Table 4 above, the second to last column shows the average CPU utilization. The rows all show the performance when doing about the same load per node (12000 operations per second). Note that the average CPU busy time per operation increases as the number of nodes increases. To account for some of the load variability an estimation of the CPU time per operation can be made by dividing the total busy microseconds (calculated from sar cpu busy) per second by the operations per second. This is what is reported in the last column of Table 4 as Approx CPU microsecond busy per operation. For example, at the operations per second per node load point, an operation took about 15% more time on the 16-node configuration as compared to the 1-node configuration. The cause of this increase was not investigated as a part of this project, but this type of increase is commonly seen in multi-processors servers as the number of CPUs is increased. One likely cause for the increase is an increase in the number of instructions with more nodes as the potential for contention and the need for coordination increase. Figure 4: Scalability factor at 80% CPU busy 13/14

14 The tables below show the throughput ratios for the various configurations using the two methods illustrated in Figures Figure 2 through Figure 4. Table 5 and Table 6 show the ratio of the throughput for the N-node configuration (far left column) divided by the throughput for the 1- node, 2-node,, 12-node configurations. The far right hand column shows the scalability as a percentage (scalability vs. 1-node divided by the number of nodes). In Table 5, the values shown are from the maximum throughputs for each test (values shown in Figure 2); in Table 6, the values are from the measured throughput tests when each configuration was about 80% CPU busy (values used in Figures Figure 3 and Figure 4). The values in Table 5 are all very high; again, these values do not factor in the fact that the throughput in the smaller configurations is limited by the array. The values in Table 6 (that factor in the CPU utilization) are still quite high and show that the 16 nodes scalability (as a percentage) is 91% for this particular workload. Table 5: Throughput Ratios of N-nodes vs. M-nodes at Maximum Throughput Number Of Nodes Vs. 1 Node vs. 2 Node vs. 4 Node vs. 8 Node vs. 12 Node Percentage of Perfect Scalability % % % % % % Table 6: Throughput Ratios of N-nodes vs. M-nodes using Throughput at 80% CPU busy Number Of Nodes Vs. 1 Node vs. 2 Node vs. 4 Node vs. 8 Node vs. 12 Node Percentage of Perfect Scalability % % % % % % 14/14

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Oracle Database Scalability in VMware ESX VMware ESX 3.5 Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises

More information

Comparing the Network Performance of Windows File Sharing Environments

Comparing the Network Performance of Windows File Sharing Environments Technical Report Comparing the Network Performance of Windows File Sharing Environments Dan Chilton, Srinivas Addanki, NetApp September 2010 TR-3869 EXECUTIVE SUMMARY This technical report presents the

More information

HP reference configuration for entry-level SAS Grid Manager solutions

HP reference configuration for entry-level SAS Grid Manager solutions HP reference configuration for entry-level SAS Grid Manager solutions Up to 864 simultaneous SAS jobs and more than 3 GB/s I/O throughput Technical white paper Table of contents Executive summary... 2

More information

Virtuoso and Database Scalability

Virtuoso and Database Scalability Virtuoso and Database Scalability By Orri Erling Table of Contents Abstract Metrics Results Transaction Throughput Initializing 40 warehouses Serial Read Test Conditions Analysis Working Set Effect of

More information

Microsoft Windows Server 2003 with Internet Information Services (IIS) 6.0 vs. Linux Competitive Web Server Performance Comparison

Microsoft Windows Server 2003 with Internet Information Services (IIS) 6.0 vs. Linux Competitive Web Server Performance Comparison April 23 11 Aviation Parkway, Suite 4 Morrisville, NC 2756 919-38-28 Fax 919-38-2899 32 B Lakeside Drive Foster City, CA 9444 65-513-8 Fax 65-513-899 www.veritest.com info@veritest.com Microsoft Windows

More information

New!! - Higher performance for Windows and UNIX environments

New!! - Higher performance for Windows and UNIX environments New!! - Higher performance for Windows and UNIX environments The IBM TotalStorage Network Attached Storage Gateway 300 (NAS Gateway 300) is designed to act as a gateway between a storage area network (SAN)

More information

Oracle Database Deployments with EMC CLARiiON AX4 Storage Systems

Oracle Database Deployments with EMC CLARiiON AX4 Storage Systems Oracle Database Deployments with EMC CLARiiON AX4 Storage Systems Applied Technology Abstract This white paper investigates configuration and replication choices for Oracle Database deployment with EMC

More information

POSIX and Object Distributed Storage Systems

POSIX and Object Distributed Storage Systems 1 POSIX and Object Distributed Storage Systems Performance Comparison Studies With Real-Life Scenarios in an Experimental Data Taking Context Leveraging OpenStack Swift & Ceph by Michael Poat, Dr. Jerome

More information

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays Database Solutions Engineering By Murali Krishnan.K Dell Product Group October 2009

More information

Violin Memory 7300 Flash Storage Platform Supports Multiple Primary Storage Workloads

Violin Memory 7300 Flash Storage Platform Supports Multiple Primary Storage Workloads Violin Memory 7300 Flash Storage Platform Supports Multiple Primary Storage Workloads Web server, SQL Server OLTP, Exchange Jetstress, and SharePoint Workloads Can Run Simultaneously on One Violin Memory

More information

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies

More information

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010 Best Practices for Data Sharing in a Grid Distributed SAS Environment Updated July 2010 B E S T P R A C T I C E D O C U M E N T Table of Contents 1 Abstract... 2 1.1 Storage performance is critical...

More information

Shared Parallel File System

Shared Parallel File System Shared Parallel File System Fangbin Liu fliu@science.uva.nl System and Network Engineering University of Amsterdam Shared Parallel File System Introduction of the project The PVFS2 parallel file system

More information

Performance and scalability of a large OLTP workload

Performance and scalability of a large OLTP workload Performance and scalability of a large OLTP workload ii Performance and scalability of a large OLTP workload Contents Performance and scalability of a large OLTP workload with DB2 9 for System z on Linux..............

More information

PERFORMANCE TUNING ORACLE RAC ON LINUX

PERFORMANCE TUNING ORACLE RAC ON LINUX PERFORMANCE TUNING ORACLE RAC ON LINUX By: Edward Whalen Performance Tuning Corporation INTRODUCTION Performance tuning is an integral part of the maintenance and administration of the Oracle database

More information

Condusiv s V-locity Server Boosts Performance of SQL Server 2012 by 55%

Condusiv s V-locity Server Boosts Performance of SQL Server 2012 by 55% openbench Labs Executive Briefing: April 19, 2013 Condusiv s Server Boosts Performance of SQL Server 2012 by 55% Optimizing I/O for Increased Throughput and Reduced Latency on Physical Servers 01 Executive

More information

Use of Hadoop File System for Nuclear Physics Analyses in STAR

Use of Hadoop File System for Nuclear Physics Analyses in STAR 1 Use of Hadoop File System for Nuclear Physics Analyses in STAR EVAN SANGALINE UC DAVIS Motivations 2 Data storage a key component of analysis requirements Transmission and storage across diverse resources

More information

AIX NFS Client Performance Improvements for Databases on NAS

AIX NFS Client Performance Improvements for Databases on NAS AIX NFS Client Performance Improvements for Databases on NAS October 20, 2005 Sanjay Gulabani Sr. Performance Engineer Network Appliance, Inc. gulabani@netapp.com Diane Flemming Advisory Software Engineer

More information

InfoScale Storage & Media Server Workloads

InfoScale Storage & Media Server Workloads InfoScale Storage & Media Server Workloads Maximise Performance when Storing and Retrieving Large Amounts of Unstructured Data Carlos Carrero Colin Eldridge Shrinivas Chandukar 1 Table of Contents 01 Introduction

More information

Microsoft Exchange Server 2003 Deployment Considerations

Microsoft Exchange Server 2003 Deployment Considerations Microsoft Exchange Server 3 Deployment Considerations for Small and Medium Businesses A Dell PowerEdge server can provide an effective platform for Microsoft Exchange Server 3. A team of Dell engineers

More information

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1 Performance Study Performance Characteristics of and RDM VMware ESX Server 3.0.1 VMware ESX Server offers three choices for managing disk access in a virtual machine VMware Virtual Machine File System

More information

PRODUCT BRIEF 3E PERFORMANCE BENCHMARKS LOAD AND SCALABILITY TESTING

PRODUCT BRIEF 3E PERFORMANCE BENCHMARKS LOAD AND SCALABILITY TESTING PRODUCT BRIEF 3E PERFORMANCE BENCHMARKS LOAD AND SCALABILITY TESTING THE FOUNDATION Thomson Reuters Elite completed a series of performance load tests with the 3E application to verify that it could scale

More information

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003

Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks. An Oracle White Paper April 2003 Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building Blocks An Oracle White Paper April 2003 Achieving Mainframe-Class Performance on Intel Servers Using InfiniBand Building

More information

Recommended hardware system configurations for ANSYS users

Recommended hardware system configurations for ANSYS users Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range

More information

Dell Reference Configuration for Hortonworks Data Platform

Dell Reference Configuration for Hortonworks Data Platform Dell Reference Configuration for Hortonworks Data Platform A Quick Reference Configuration Guide Armando Acosta Hadoop Product Manager Dell Revolutionary Cloud and Big Data Group Kris Applegate Solution

More information

PARALLELS CLOUD STORAGE

PARALLELS CLOUD STORAGE PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...

More information

Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array

Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array Evaluation report prepared under contract with Lenovo Executive Summary Even with the price of flash

More information

Maximum performance, minimal risk for data warehousing

Maximum performance, minimal risk for data warehousing SYSTEM X SERVERS SOLUTION BRIEF Maximum performance, minimal risk for data warehousing Microsoft Data Warehouse Fast Track for SQL Server 2014 on System x3850 X6 (95TB) The rapid growth of technology has

More information

EMC Virtual Infrastructure for Microsoft SQL Server

EMC Virtual Infrastructure for Microsoft SQL Server Microsoft SQL Server Enabled by EMC Celerra and Microsoft Hyper-V Copyright 2010 EMC Corporation. All rights reserved. Published February, 2010 EMC believes the information in this publication is accurate

More information

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC

More information

MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM?

MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? Ashutosh Shinde Performance Architect ashutosh_shinde@hotmail.com Validating if the workload generated by the load generating tools is applied

More information

Operating System for the K computer

Operating System for the K computer Operating System for the K computer Jun Moroo Masahiko Yamada Takeharu Kato For the K computer to achieve the world s highest performance, Fujitsu has worked on the following three performance improvements

More information

Sun 8Gb/s Fibre Channel HBA Performance Advantages for Oracle Database

Sun 8Gb/s Fibre Channel HBA Performance Advantages for Oracle Database Performance Advantages for Oracle Database At a Glance This Technical Brief illustrates that even for smaller online transaction processing (OLTP) databases, the Sun 8Gb/s Fibre Channel Host Bus Adapter

More information

Best practices for Implementing Lotus Domino in a Storage Area Network (SAN) Environment

Best practices for Implementing Lotus Domino in a Storage Area Network (SAN) Environment Best practices for Implementing Lotus Domino in a Storage Area Network (SAN) Environment With the implementation of storage area networks (SAN) becoming more of a standard configuration, this paper describes

More information

Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat

Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat Why Computers Are Getting Slower The traditional approach better performance Why computers are

More information

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000 The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000 Summary: This document describes how to analyze performance on an IBM Storwize V7000. IntelliMagic 2012 Page 1 This

More information

Benchmarking Cassandra on Violin

Benchmarking Cassandra on Violin Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract

More information

Storage benchmarking cookbook

Storage benchmarking cookbook Storage benchmarking cookbook How to perform solid storage performance measurements Stijn Eeckhaut Stijn De Smet, Brecht Vermeulen, Piet Demeester The situation today: storage systems can be very complex

More information

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study White Paper Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study 2012 Cisco and/or its affiliates. All rights reserved. This

More information

Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009

Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Performance Study Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Introduction With more and more mission critical networking intensive workloads being virtualized

More information

SQL Server Business Intelligence on HP ProLiant DL785 Server

SQL Server Business Intelligence on HP ProLiant DL785 Server SQL Server Business Intelligence on HP ProLiant DL785 Server By Ajay Goyal www.scalabilityexperts.com Mike Fitzner Hewlett Packard www.hp.com Recommendations presented in this document should be thoroughly

More information

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V Copyright 2011 EMC Corporation. All rights reserved. Published February, 2011 EMC believes the information

More information

IP SAN Fundamentals: An Introduction to IP SANs and iscsi

IP SAN Fundamentals: An Introduction to IP SANs and iscsi IP SAN Fundamentals: An Introduction to IP SANs and iscsi Updated April 2007 Sun Microsystems, Inc. 2007 Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, CA 95054 USA All rights reserved. This

More information

Diablo and VMware TM powering SQL Server TM in Virtual SAN TM. A Diablo Technologies Whitepaper. May 2015

Diablo and VMware TM powering SQL Server TM in Virtual SAN TM. A Diablo Technologies Whitepaper. May 2015 A Diablo Technologies Whitepaper Diablo and VMware TM powering SQL Server TM in Virtual SAN TM May 2015 Ricky Trigalo, Director for Virtualization Solutions Architecture, Diablo Technologies Daniel Beveridge,

More information

EMC Unified Storage for Microsoft SQL Server 2008

EMC Unified Storage for Microsoft SQL Server 2008 EMC Unified Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON and EMC FAST Cache Reference Copyright 2010 EMC Corporation. All rights reserved. Published October, 2010 EMC believes the information

More information

Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering

Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations A Dell Technical White Paper Database Solutions Engineering By Sudhansu Sekhar and Raghunatha

More information

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance. Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance

More information

Delivering Quality in Software Performance and Scalability Testing

Delivering Quality in Software Performance and Scalability Testing Delivering Quality in Software Performance and Scalability Testing Abstract Khun Ban, Robert Scott, Kingsum Chow, and Huijun Yan Software and Services Group, Intel Corporation {khun.ban, robert.l.scott,

More information

Seradex White Paper. Focus on these points for optimizing the performance of a Seradex ERP SQL database:

Seradex White Paper. Focus on these points for optimizing the performance of a Seradex ERP SQL database: Seradex White Paper A Discussion of Issues in the Manufacturing OrderStream Microsoft SQL Server High Performance for Your Business Executive Summary Microsoft SQL Server is the leading database product

More information

Cray DVS: Data Virtualization Service

Cray DVS: Data Virtualization Service Cray : Data Virtualization Service Stephen Sugiyama and David Wallace, Cray Inc. ABSTRACT: Cray, the Cray Data Virtualization Service, is a new capability being added to the XT software environment with

More information

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

HP SN1000E 16 Gb Fibre Channel HBA Evaluation HP SN1000E 16 Gb Fibre Channel HBA Evaluation Evaluation report prepared under contract with Emulex Executive Summary The computing industry is experiencing an increasing demand for storage performance

More information

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup

More information

Deployments and Tests in an iscsi SAN

Deployments and Tests in an iscsi SAN Deployments and Tests in an iscsi SAN SQL Server Technical Article Writer: Jerome Halmans, Microsoft Corp. Technical Reviewers: Eric Schott, EqualLogic, Inc. Kevin Farlee, Microsoft Corp. Darren Miller,

More information

Post-production Video Editing Solution Guide with Quantum StorNext File System AssuredSAN 4000

Post-production Video Editing Solution Guide with Quantum StorNext File System AssuredSAN 4000 Post-production Video Editing Solution Guide with Quantum StorNext File System AssuredSAN 4000 Dot Hill Systems introduction 1 INTRODUCTION Dot Hill Systems offers high performance network storage products

More information

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2008 R2 Hyper-V Live Migration Windows Server 2008 R2 Hyper-V Live Migration Table of Contents Overview of Windows Server 2008 R2 Hyper-V Features... 3 Dynamic VM storage... 3 Enhanced Processor Support... 3 Enhanced Networking Support...

More information

Maximizing VMware ESX Performance Through Defragmentation of Guest Systems. Presented by

Maximizing VMware ESX Performance Through Defragmentation of Guest Systems. Presented by Maximizing VMware ESX Performance Through Defragmentation of Guest Systems Presented by July, 2010 Table of Contents EXECUTIVE OVERVIEW 3 TEST EQUIPMENT AND METHODS 4 TESTING OVERVIEW 5 Fragmentation in

More information

Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage

Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage Technical white paper Table of contents Executive summary... 2 Introduction... 2 Test methodology... 3

More information

Boost Database Performance with the Cisco UCS Storage Accelerator

Boost Database Performance with the Cisco UCS Storage Accelerator Boost Database Performance with the Cisco UCS Storage Accelerator Performance Brief February 213 Highlights Industry-leading Performance and Scalability Offloading full or partial database structures to

More information

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator

More information

Online Remote Data Backup for iscsi-based Storage Systems

Online Remote Data Backup for iscsi-based Storage Systems Online Remote Data Backup for iscsi-based Storage Systems Dan Zhou, Li Ou, Xubin (Ben) He Department of Electrical and Computer Engineering Tennessee Technological University Cookeville, TN 38505, USA

More information

Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers

Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers WHITE PAPER FUJITSU PRIMERGY AND PRIMEPOWER SERVERS Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers CHALLENGE Replace a Fujitsu PRIMEPOWER 2500 partition with a lower cost solution that

More information

D1.2 Network Load Balancing

D1.2 Network Load Balancing D1. Network Load Balancing Ronald van der Pol, Freek Dijkstra, Igor Idziejczak, and Mark Meijerink SARA Computing and Networking Services, Science Park 11, 9 XG Amsterdam, The Netherlands June ronald.vanderpol@sara.nl,freek.dijkstra@sara.nl,

More information

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing An Alternative Storage Solution for MapReduce Eric Lomascolo Director, Solutions Marketing MapReduce Breaks the Problem Down Data Analysis Distributes processing work (Map) across compute nodes and accumulates

More information

Esri ArcGIS Server 10 for VMware Infrastructure

Esri ArcGIS Server 10 for VMware Infrastructure Esri ArcGIS Server 10 for VMware Infrastructure October 2011 DEPLOYMENT AND TECHNICAL CONSIDERATIONS GUIDE Table of Contents Introduction... 3 Esri ArcGIS Server 10 Overview.... 3 VMware Infrastructure

More information

PARALLELS CLOUD SERVER

PARALLELS CLOUD SERVER PARALLELS CLOUD SERVER Performance and Scalability 1 Table of Contents Executive Summary... Error! Bookmark not defined. LAMP Stack Performance Evaluation... Error! Bookmark not defined. Background...

More information

The IntelliMagic White Paper on: Storage Performance Analysis for an IBM San Volume Controller (SVC) (IBM V7000)

The IntelliMagic White Paper on: Storage Performance Analysis for an IBM San Volume Controller (SVC) (IBM V7000) The IntelliMagic White Paper on: Storage Performance Analysis for an IBM San Volume Controller (SVC) (IBM V7000) IntelliMagic, Inc. 558 Silicon Drive Ste 101 Southlake, Texas 76092 USA Tel: 214-432-7920

More information

Benchmarking FreeBSD. Ivan Voras <ivoras@freebsd.org>

Benchmarking FreeBSD. Ivan Voras <ivoras@freebsd.org> Benchmarking FreeBSD Ivan Voras What and why? Everyone likes a nice benchmark graph :) And it's nice to keep track of these things The previous major run comparing FreeBSD to Linux

More information

SCI Briefing: A Review of the New Hitachi Unified Storage and Hitachi NAS Platform 4000 Series. Silverton Consulting, Inc.

SCI Briefing: A Review of the New Hitachi Unified Storage and Hitachi NAS Platform 4000 Series. Silverton Consulting, Inc. SCI Briefing: A Review of the New Hitachi Unified Storage and Hitachi NAS Platform 4000 Series Silverton Consulting, Inc. StorInt Briefing Written by: Ray Lucchesi, President and Founder Published: July,

More information

WHITE PAPER BRENT WELCH NOVEMBER

WHITE PAPER BRENT WELCH NOVEMBER BACKUP WHITE PAPER BRENT WELCH NOVEMBER 2006 WHITE PAPER: BACKUP TABLE OF CONTENTS Backup Overview 3 Background on Backup Applications 3 Backup Illustration 4 Media Agents & Keeping Tape Drives Busy 5

More information

Oracle TimesTen In-Memory Database on Oracle Exalogic Elastic Cloud

Oracle TimesTen In-Memory Database on Oracle Exalogic Elastic Cloud An Oracle White Paper July 2011 Oracle TimesTen In-Memory Database on Oracle Exalogic Elastic Cloud Executive Summary... 3 Introduction... 4 Hardware and Software Overview... 5 Compute Node... 5 Storage

More information

Oracle9i Release 2 Database Architecture on Windows. An Oracle Technical White Paper April 2003

Oracle9i Release 2 Database Architecture on Windows. An Oracle Technical White Paper April 2003 Oracle9i Release 2 Database Architecture on Windows An Oracle Technical White Paper April 2003 Oracle9i Release 2 Database Architecture on Windows Executive Overview... 3 Introduction... 3 Oracle9i Release

More information

VMware VMmark V1.1.1 Results

VMware VMmark V1.1.1 Results VMware VMmark V1.1.1 Results Vendor and Hardware Platform: Lenovo R525 G2 Virtualization Platform: VMware ESX 4.0(build 164009) Performance Section Performance Tested By: Lenovo Test Date: 06/05/2009 Configuration

More information

Adaptec: Snap Server NAS Performance Study

Adaptec: Snap Server NAS Performance Study March 2006 www.veritest.com info@veritest.com Adaptec: Snap Server NAS Performance Study Test report prepared under contract from Adaptec, Inc. Executive summary Adaptec commissioned VeriTest, a service

More information

Scalability Factors of JMeter In Performance Testing Projects

Scalability Factors of JMeter In Performance Testing Projects Scalability Factors of JMeter In Performance Testing Projects Title Scalability Factors for JMeter In Performance Testing Projects Conference STEP-IN Conference Performance Testing 2008, PUNE Author(s)

More information

Dell EqualLogic Best Practices Series

Dell EqualLogic Best Practices Series Dell EqualLogic Best Practices Series Sizing and Best Practices for Deploying Oracle 11g Release 2 Based Decision Support Systems with Dell EqualLogic 10GbE iscsi SAN A Dell Technical Whitepaper Storage

More information

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre University of Cambridge, UIS, HPC Service Authors: Wojciech Turek, Paul Calleja, John Taylor

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

Best Practices for Optimizing Storage for Oracle Automatic Storage Management with Oracle FS1 Series Storage ORACLE WHITE PAPER JANUARY 2015

Best Practices for Optimizing Storage for Oracle Automatic Storage Management with Oracle FS1 Series Storage ORACLE WHITE PAPER JANUARY 2015 Best Practices for Optimizing Storage for Oracle Automatic Storage Management with Oracle FS1 Series Storage ORACLE WHITE PAPER JANUARY 2015 Table of Contents 0 Introduction 1 The Test Environment 1 Best

More information

W H I T E P A P E R. Performance and Scalability of Microsoft SQL Server on VMware vsphere 4

W H I T E P A P E R. Performance and Scalability of Microsoft SQL Server on VMware vsphere 4 W H I T E P A P E R Performance and Scalability of Microsoft SQL Server on VMware vsphere 4 Table of Contents Introduction................................................................... 3 Highlights.....................................................................

More information

Windows Server 2008 R2 Hyper-V Live Migration

Windows Server 2008 R2 Hyper-V Live Migration Windows Server 2008 R2 Hyper-V Live Migration White Paper Published: August 09 This is a preliminary document and may be changed substantially prior to final commercial release of the software described

More information

Quantum StorNext. Product Brief: Distributed LAN Client

Quantum StorNext. Product Brief: Distributed LAN Client Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without

More information

Accelerating and Simplifying Apache

Accelerating and Simplifying Apache Accelerating and Simplifying Apache Hadoop with Panasas ActiveStor White paper NOvember 2012 1.888.PANASAS www.panasas.com Executive Overview The technology requirements for big data vary significantly

More information

Improved LS-DYNA Performance on Sun Servers

Improved LS-DYNA Performance on Sun Servers 8 th International LS-DYNA Users Conference Computing / Code Tech (2) Improved LS-DYNA Performance on Sun Servers Youn-Seo Roh, Ph.D. And Henry H. Fong Sun Microsystems, Inc. Abstract Current Sun platforms

More information

The functionality and advantages of a high-availability file server system

The functionality and advantages of a high-availability file server system The functionality and advantages of a high-availability file server system This paper discusses the benefits of deploying a JMR SHARE High-Availability File Server System. Hardware and performance considerations

More information

Tableau Server Scalability Explained

Tableau Server Scalability Explained Tableau Server Scalability Explained Author: Neelesh Kamkolkar Tableau Software July 2013 p2 Executive Summary In March 2013, we ran scalability tests to understand the scalability of Tableau 8.0. We wanted

More information

Enabling Technologies for Distributed Computing

Enabling Technologies for Distributed Computing Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing, UNF Multi-core CPUs and Multithreading Technologies

More information

SIDN Server Measurements

SIDN Server Measurements SIDN Server Measurements Yuri Schaeffer 1, NLnet Labs NLnet Labs document 2010-003 July 19, 2010 1 Introduction For future capacity planning SIDN would like to have an insight on the required resources

More information

Virtualizing Microsoft SQL Server 2008 on the Hitachi Adaptable Modular Storage 2000 Family Using Microsoft Hyper-V

Virtualizing Microsoft SQL Server 2008 on the Hitachi Adaptable Modular Storage 2000 Family Using Microsoft Hyper-V Virtualizing Microsoft SQL Server 2008 on the Hitachi Adaptable Modular Storage 2000 Family Using Microsoft Hyper-V Implementation Guide By Eduardo Freitas and Ryan Sokolowski February 2010 Summary Deploying

More information

White Paper. Recording Server Virtualization

White Paper. Recording Server Virtualization White Paper Recording Server Virtualization Prepared by: Mike Sherwood, Senior Solutions Engineer Milestone Systems 23 March 2011 Table of Contents Introduction... 3 Target audience and white paper purpose...

More information

Optimizing LTO Backup Performance

Optimizing LTO Backup Performance Optimizing LTO Backup Performance July 19, 2011 Written by: Ash McCarty Contributors: Cedrick Burton Bob Dawson Vang Nguyen Richard Snook Table of Contents 1.0 Introduction... 3 2.0 Host System Configuration...

More information

Performance And Scalability In Oracle9i And SQL Server 2000

Performance And Scalability In Oracle9i And SQL Server 2000 Performance And Scalability In Oracle9i And SQL Server 2000 Presented By : Phathisile Sibanda Supervisor : John Ebden 1 Presentation Overview Project Objectives Motivation -Why performance & Scalability

More information

Accelerating Microsoft Exchange Servers with I/O Caching

Accelerating Microsoft Exchange Servers with I/O Caching Accelerating Microsoft Exchange Servers with I/O Caching QLogic FabricCache Caching Technology Designed for High-Performance Microsoft Exchange Servers Key Findings The QLogic FabricCache 10000 Series

More information

Oracle Applications Release 10.7 NCA Network Performance for the Enterprise. An Oracle White Paper January 1998

Oracle Applications Release 10.7 NCA Network Performance for the Enterprise. An Oracle White Paper January 1998 Oracle Applications Release 10.7 NCA Network Performance for the Enterprise An Oracle White Paper January 1998 INTRODUCTION Oracle has quickly integrated web technologies into business applications, becoming

More information

HP ProLiant DL585 G5 earns #1 virtualization performance record on VMmark Benchmark

HP ProLiant DL585 G5 earns #1 virtualization performance record on VMmark Benchmark HP ProLiant DL585 G5 earns #1 virtualization performance record on VMmark Benchmark HP Leadership»The HP ProLiant DL585 G5 is a highly manageable, rack optimized, four-socket server designed for maximum

More information

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Agenda Introduction Database Architecture Direct NFS Client NFS Server

More information

EMC Backup and Recovery for Microsoft Exchange 2007 SP2

EMC Backup and Recovery for Microsoft Exchange 2007 SP2 EMC Backup and Recovery for Microsoft Exchange 2007 SP2 Enabled by EMC Celerra and Microsoft Windows 2008 Copyright 2010 EMC Corporation. All rights reserved. Published February, 2010 EMC believes the

More information

Managing Storage Space in a Flash and Disk Hybrid Storage System

Managing Storage Space in a Flash and Disk Hybrid Storage System Managing Storage Space in a Flash and Disk Hybrid Storage System Xiaojian Wu, and A. L. Narasimha Reddy Dept. of Electrical and Computer Engineering Texas A&M University IEEE International Symposium on

More information

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments 8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments QLogic 8Gb Adapter Outperforms Emulex QLogic Offers Best Performance and Scalability in Hyper-V Environments Key Findings The QLogic

More information

Software-defined Storage Architecture for Analytics Computing

Software-defined Storage Architecture for Analytics Computing Software-defined Storage Architecture for Analytics Computing Arati Joshi Performance Engineering Colin Eldridge File System Engineering Carlos Carrero Product Management June 2015 Reference Architecture

More information