A Survey of Shared File Systems

Size: px
Start display at page:

Download "A Survey of Shared File Systems"

Transcription

1 Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications Updated 22 Oct 2014

2 A Survey of Shared File Systems

3 A Survey of Shared File Systems Table of Contents Introduction... 1 Retaining File System Data in Memory... 1 Maintenance of File System Metadata... 1 Implications for Physical Resources... 2 Workloads... 2 Large Sorts... 2 Demand Model Forecasting and Calibration... 2 Configuration... 3 File System Configuration... 3 SAS System Options... 3 IBM Elastic Storage (aka General Parallel File System or GPFS )... 5 IBM General Parallel File System File Placement Optimizer (GPFS FPO). 6 NFS... 7 Global File System2 (Red Hat Resilient Storage)... 8 Common Internet File System (CIFS)... 9 Quantum StorNext Intel Enterprise Edition for Lustre Veritas Cluster File System... 12

4 A Survey of Shared File Systems

5 A Survey of Shared File Systems 1 Introduction A shared file system is a required and integral component of all SAS Grid Manager deployments, Enterprise Business Intelligence deployments with load balanced servers on multiple systems, and other types of distributed SAS applications. In order to determine which shared file system is the best choice for a given deployment, it is important to understand how SAS software interacts with the file system and how a shared file system might behave differently compared to a nonshared file system. The important characteristics of shared file systems with respect to SAS performance are whether the file system data is retained in memory in a local file cache handling of file system metadata implications for the physical resources This paper briefly describes these aspects and examines the behavior of several different shared file systems in the context of performance of a representative deployment. Retaining File System Data in Memory Data read from storage such as a solid state device or disk is referred to as a physical read. When the data is retained in and read from the local file cache the reads are called logical reads. Programs that perform logical reads from data that has already been read physically into memory perform substantially better than when the program performs physical reads. SAS programs often read the same data multiple times. If the data are retained in a local file cache, the program can perform logical reads from the local file cache much faster and more efficiently than re-reading it from the physical device. Operations that access small files tend to get a substantial performance benefit from the local file cache. Files that are larger than the size of the cache memory cannot be cached in their entirety and may not be cached at all. Most non-shared file systems will retain data in the local file cache until the memory is needed for some other purpose. In comparison, shared file systems implement a wide variety of rules that govern the retention of data in the local file cache. Some file systems allow tuning of the local file cache, providing options for limiting the amount of memory that can be used for file cache or the number of files kept in the local file cache, for example. Maintenance of File System Metadata File system metadata includes such information as lists of files in a directory, file attributes such as file permissions or file creation date, and other information about the physical data. Various shared file systems differ in the maintenance of the file system metadata. Shared file systems are required to make information about file system metadata and file locking available to all systems participating in the shared file system. File system metadata is updated whenever a file is created, modified, deleted or extended, when a lock is obtained or dropped, and on some file systems, when a file is accessed. The sizes of files created by SAS programs are not known when files are first created. As the file is written, additional file system blocks, or extents, are requested. Each extent request results in an update to file system metadata. In a shared file

6 2 A Survey of Shared File Systems system, file system metadata changes are coordinated across systems, so there is more overhead associated with file system metadata with a shared file system compared to a non-shared file system. Due to the differences in local file cache behavior and the additional overhead required to maintain file system metadata, programs that performed well with a non-shared file system may behave quite differently with a shared file system. To minimize overhead from the shared file system, a general recommendation is to maintain SAS WORK directories in a nonshared file system and place only permanent data in a shared file system. A shared file system may be necessary for GRIDWORK. SAS WORK files are not usually shared and are considered to be private to the process that created the directory, so there is no benefit for SAS WORK files to reside in a shared file system. SAS WORK files also tend to generate a higher number of file system metadata updates because files are continually created, extended and deleted which can greatly increase the overhead of coordinating file system metadata in the shared file system. In addition, a non-shared file system is more likely to retain SAS WORK files in the local file cache. Implications for Physical Resources Shared file systems require coordination between multiple host systems and place more demand on the network and storage devices. This means that the network and storage devices must be provisioned appropriately. In addition, less efficient file cache management may result in more data transfer and more metadata requests, both of which may result in more physical I/O. The storage subsystem may need to support both a higher number of I/O operations as well as a higher throughput rate compared to the needs of a non-shared file system. Workloads In order to characterize different shared file systems and storage devices, we developed a small set of internal benchmarks to exercise typical SAS software scenarios. The following is a description of two of the workloads that are being used to study a variety of shared file systems. The outcomes for each of those shared file systems are described later in the paper. Large Sorts Large sorting workloads measure throughput of the file/storage subsystem. This workload is intended to push the I/O subsystem bandwidth to its limit. The workload launches multiple simultaneous sorts, typically 1 to 2 per CPU core. The files used by this workload are typically larger than system memory and will not be retained in the local file cache. File system metadata is exercised lightly. Demand Model Forecasting and Calibration This workload runs a configurable number of concurrent SAS processes that execute a demand model calibration process. The workload emulates a SAS solution deployed in a distributed environment using shared input data and a shared output analytic data mart. The SAS programs are written using SAS macro language and create many small to moderate datasets.

7 A Survey of Shared File Systems 3 There are 750 models in the workload and each model is calibrated by an individual SAS process. Each of these processes creates about 100 permanent output datasets and 7,000 temporary files. Many solutions, including SAS Drug Development, SAS Warranty Analysis, SAS Enterprise Miner and others behave in a similar manner. This workload will exercise the management of file system metadata and some shared file systems have failed with what we consider a representative load. This workload tends to benefit a great deal from the file cache, because there are many small files that are read multiple times. Configuration It is important to configure your system appropriately before running SAS applications. Below we discuss file system and SAS options used for our shared file system studies. File System Configuration Some shared file systems are available as NAS (network attached storage) rather than SAN (storage area network). With 10 GbE and other high bandwidth network connectivity, NAS devices have become much more attractive because much higher throughput rates can be achieved compared to lower bandwidth connectivity such as 1 GbE. However, even with the higher bandwidth networks, throughput may not be sufficient for systems with a high core count. Systems with 12 or more cores may not be fully utilized due to constraints and latency of the network. There are some file system options that are available on most Unix-based or Windows-based systems that enable or disable recording when a file is accessed. Unless the workload specifically requires information about when a file was accessed or modified, disabling the metadata update with shared file systems reduces some of the load on the file system. On Unix-based systems, mounting file systems with the option NOATIME= disables the recording of file access times. The mount option NOATIME= improved the performance of all benchmarks and is a recommended setting if accurate file access time is not required for correct program execution. On Microsoft Windows operating systems, setting the registry value HKLM\System\CurrentControlSet\Control\Filesystem\NtfsDisableLastAccessUpdate to 1 also disables this type of file system metadata update. SAS System Options Where possible, the SAS options are set to coordinate well with the file system page size and device stripe size. The option ALIGNSASIOFILES aligns all datasets, including SAS WORK datasets on the specified boundary. It performs the same alignment as the LIBNAME option OFFSET specified for non- WORK libraries. Other options align SAS I/O requests to the boundary specified by the buffer size. I/O alignment may reduce the number of physical I/O operations. For shared file systems, these options are particularly useful because many shared file systems will not coalesce unaligned I/O requests. Misalignment may cause a much higher number of requests, either over a network or to the I/O device, compared to aligned requests. The following is a description of the SAS Foundation 9.3 options used to tune performance of the benchmarks: CPUCOUNT=2

8 4 A Survey of Shared File Systems These benchmarks all ran multiple SAS processes concurrently. The option CPUCOUNT was set to 2 in order to minimize contention for CPU resources. MEMSIZE=1G This option MEMSIZE was set to 1G minimize contention for memory resources. SORTSIZE=768m The option SORTSIZE was set to 75% of memsize. ALIGNSASIOFILES The option ALIGNSASIOFILES aligns access to SAS datasets on a specific boundary. The boundary is defined by BUFSIZE. This allows a file system to align read and write requests and avoid split I/O operations. BUFSIZE=128k The option BUFSIZE sets the read/write size of each request made to SAS datasets. UBUFSIZE=128k The option UBUFSIZE sets the read/write size of SAS utility file requests. IBUFSIZE=32767 The option IBUFSIZE sets the read/write size of SAS index file requests. IBUFNO=10 Some of the workloads utilize indexes. Setting IBUFNO to 10 reduced the number of file system requests for indexed traversal. FILELOCKS=none By default, on UNIX systems the SAS option FILELOCKS=FAIL is set. This setting obtains read and write file locks on non- WORK files to protect programs from making concurrent writes to a file or from reading data that may be in an inconsistent state. File systems make metadata requests in order to manage locks. The benchmarks were run with both FILELOCKS=FAIL and FILELOCKS=NONE in order to determine the overhead associated with file locks. In order to reduce metadata requests, there are SAS options that eliminate file lock requests. For data that is not shared, is only read, or if the application is designed not to need file locks, setting the option FILELOCKS=NONE will reduce file system metadata requests. Also, some file systems will flush files from memory when any file attribute changes. This means files accessed with a lock may be forced out of the local file cache and re-read from storage. The overhead from managing file locks and the change in file cache retention can be substantial, therefore limiting the use of file locks to just those files and libraries that require concurrent read/write access is recommended for best performance. File locks are necessary to avoid corrupting data or accessing data in an inconsistent state. Examples of situations that require the use of file locks are: concurrent processes attempting to update the same file a single process updating a file and other processes attempting to read it concurrently.

9 A Survey of Shared File Systems 5 We are in the process of running these workloads on a variety of shared file systems. The following sections describe the file systems benchmarked as of April, File System Performance Studies In this section we discuss the shared file systems studied to date. We plan to study additional file systems and report those results also. IBM Elastic Storage (aka General Parallel File System or GPFS ) As of the release of version 4.1 IBM is rebranding GPFS as IBM Elastic Storage. For the purposes of this paper IBM Elastic Storage and GPFS are synonymous. Working together SAS and IBM have found that due to certain bugs fixes and performance improvements SAS-based deployments benefit from being at IBM GPFS version or later. We advise installing at or moving to that release level or higher when possible. The IBM General Parallel File System (GPFS ) performed well on both Red Hat Enterprise Linux (RHEL) and Microsoft Windows operating systems. Both permanent data and SAS WORK files were managed by GPFS with excellent throughput and low overhead for file system metadata management. GPFS requires dedicated system memory and the amount of memory is defined by the file system option PAGEPOOL=. The client systems used in the benchmarks have 96 GB of memory and 32 GB of it was dedicated to the page pool. It is likely that client computer systems that utilize both a non-shared file system and GPFS will require additional memory to allow both file systems to perform adequately. The workloads were run with several GPFS configurations on RHEL 5.7 and 6.1 and Windows Server 2008 R2. Also, GPFS is deployed by many customers with SAS software on AIX. For RHEL, two configurations were run: a NAS configuration and a SAN configuration. The NAS configuration used a DataDirect Networks SFA10K-E GridScaler Appliance connected via 10 GbE with multiple GPFS clients running on RHEL 6.2 and Windows Server 2008 R2. The SAN configuration had SAS software running on the same systems as the GPFS NSD (network shared disk) servers using fibre channel connections to a storage device. Both configurations performed well. There was no measureable overhead using the default setting FILELOCKS=FAIL. For both the SAN and NAS configurations, GPFS was able to transmit data at a rate limited only by the throughput rate of the physical device. This rate was sustained during the large sort workload and achieved periodically for the calibration workload. The calibration workload ran 100 to 250 concurrent SAS processes spread over six 12-core systems. Maximum throughput was achieved and sustained and file system metadata overhead stayed at a manageable and responsive level. SAS data access tends to be sequential. The GPFS file system was created with the "cluster" allocation method so that data blocks would be allocated close to one another. The documentation for GPFS specifies that cluster allocation should be used for moderate configurations, but it behaved much better for both workloads compared to the "scatter" allocation method. GPFS allows separation of file system metadata from file system data. For SAS software, the access pattern to file system metadata is dramatically different than the access pattern for data. File system metadata tends to be random access of small blocks and file system data access tends to be sequential access of larger blocks. We found that creating separate volumes for file system metadata and using Solid State Devices (SSD's) provided a significant improvement in the responsiveness of the file system. The file system was created with a block size of 256 KB. Because the SAS system options BUFSIZE=128k and ALIGNSASIOFILES were specified, each dataset runs a minimum of 256 KB with the first block containing SAS header information and the next block containing data.

10 6 A Survey of Shared File Systems The following GPFS options were changed from the default values: maxmbps= The GPFS option MAXMBPS= was set to approximately twice the bandwidth available to any single system. For example, for the systems connected via 10 GbE, the value was set to 2000 (2 GB/sec). pagepool= The SAS software and all data resided in GPFS, so system memory was primarily used for either SAS processes or GPFS file cache. The GPFS option PAGEPOOL= determines the size of the file cache and was set to 32 GB. seqdiscardthreshold= The GPFS option SEQDISCARDTHRESHOLD= limits the size of a file that is read sequentially and will be retained in the file cache. This option was set to (approximately 4 GB). This value was chosen because the input data for the calibration workload is shared among many processes. If one process has already read the file, other processes can consume the same data without requiring physical I/O. However, if processes are not sharing data, a smaller value is recommended so that a few large files will not dominate the file cache. prefetchpct= Because SAS software usually reads and writes files sequentially, we increased the GPFS option PREFETCHPCT= from the default 20 percent to 40 percent. This option controls the amount of the page pool that is used for read-ahead and write-behind operations. maxfilestocache= The calibration workload creates a high number of relatively small SAS WORK files. The GPFS option MAXFILESTOCACHE= was increased from the default 1000 to The system should be monitored using a program like MMPMON to determine the appropriate value for this option. Overall, this file system behaved very well, with both the file cache benefiting performance and little overhead for maintaining file system metadata. The measured workloads had both permanent and SAS WORK files managed by GPFS. For a more detailed treatment of IBM Elastic Storage see SAS Grid Manager I/O: Optimizing SAS Application Data Availability for the Grid ( IBM General Parallel File System File Placement Optimizer (GPFS FPO) IBM introduced the File Placement Optimizer (FPO) functionality with GPFS release 3.5. FPO is not a product distinct from Elastic Storage / GPFS, rather as of version 3.5 is an included feature that supports the ability to aggregate storage across multiple servers into one or more logical file systems across a distributed shared-nothing architecture. Use of FPO brings along all the benefits of GPFS and additionally provides (1) a favorable licensing model and (2) the ability to deploy SAS Grid Manager in a shared nothing architecture reducing the need for expensive enterprise-class SAN infrastructure. GPFS FPO has the ability to aggregate storage across multiple servers into one or more logical file systems across a distributed shared-nothing architecture. GPFS FPO provides block level replication at the file system level allowing continuity of operations in light of a failed disk drive or a failed node. GPFS FPO extends the flexibility of GPFS and has

11 A Survey of Shared File Systems 7 configuration options that provides fine control over the level of replication. FPO also allows a level of control over how the initial and replica blocks are written across the cluster. For example, data written to SAS WORK will be read by the same node where it was written therefore having the ability to specify that the first block is written to the requesting node benefits SAS performance. There are two additional performance-related concerns with FPO beyond those of GPFS without use of the FPO option: Ensure each node has a sufficient quantity of disks to meet SAS recommended IO throughput requirements based upon the number of cores + overhead needed to write replica blocks from other nodes in the environment Ensure that sufficient connectivity exists that replica blocks can be written to other nodes without throttling IO throughput. SAS has found that when coupled with properly provisioned hardware GPFS FPO successfully meets the IO demands of a SAS Grid Manager implementation. For a more robust treatment of GPFS FPO functionality and SAS please see the SAS Grid Manager A Building Block approach white paper available on the Grid Computing Papers section of the Scalability and Performance focus area of support.sas.com. NFS The needs of SAS Grid Manger customers vary widely based upon workload, budget, processing windows, etc. Because of these variances it is not possible to make a blanket best practice recommendation for shared file system. NFS behaviors described in this document make the choice of NFS less desirable in demanding implementations however there are SAS customers who use NFS to their satisfaction. Certainly the use of well-engineered and properly sized NFS appliances for SASDATA is an option used by customers and for some implementations within SAS. The use of NFS for SASWORK where performance on the list of concerns is strongly discouraged due to the number of customers who have found this configuration unacceptable for performance reasons. NFS client and server implementations show a wide variety of behavior that affect performance. For that reason, if a customer chooses to use NFS - the specific client and server should be measured to ensure performance goals can be met. The performance and scalability of NFS are limited by the NFS protocol itself. If a customer implements with NFS and subsequently finds the performance does not meet business needs there is little that SAS can do to help. In this situation the customer will need to migrate to a shared file system that better aligns with their business needs. The NFS client maintains a cache of file and directory attributes. The default settings will not ensure that files created or modified on one system will be visible on another system within a minute of file creation or modification. The default settings may cause software to malfunction if multiple computer systems are accessing data that is created or modified on other computer systems. For example, if a workspace server on system A creates a new SAS dataset, that file may not be visible on System B within one minute of its creation. In order to assure a consistent view of the file system, the mount option ACTIMEO= (attribute cache timeout) should be set to 0. This setting will increase the number of requests to the NFS server for directory and file attribute information, but will ensure that the NFS client systems have a consistent view of the file system. File data modifications may not be visible on any NFS client system other than the one where the modifications are being made until an NFS commit is executed. Most NFS clients will issue a commit as part of closing a file. If multiple systems are reading files that may be modified, file system locking should be used. This is controlled by the SAS system option FILELOCKS=, the SAS library option FILELOCKS=, or the SAS LOCK statement. In order to ensure file data consistency, when an NFS client detects a change in a file system attribute, any data in the local file cache is invalidated. The next time the file is accessed, its data will be retrieved from the NFS server. This means that retention in the file cache may have much different behavior with an NFS file system compared to other file systems and the file system storage devices and network must be provisioned to handle a larger demand compared to either a local file system or a shared file system that uses a different strategy for cache coherency. Some NFS clients treat locks as file attributes. Therefore, obtaining any lock including a read/share lock may invalidate

12 8 A Survey of Shared File Systems any data that is in the local file cache and force the NFS client to request file data from the server. On UNIX systems, files in the SAS WORK directory are accessed without file locks, because those files are assumed to be private to a process. By default, the SAS system requests read and write file locks for all other files. For best performance, it is important to use locks only when necessary in order to avoid invalidating the file cache and making additional file data requests of the NFS server. File locks are needed when data can be updated by multiple concurrent processes or updated by one process and read by other concurrent processes. File locks prevent data corruption or processes reading data in an inconsistent state. The usage of file locks with NFS may have a much greater effect on overall system performance compared to other file systems. This behavior should be considered when sizing the I/O subsystem. NFS performed reasonably well on RHEL 5.7 and 6.1 in a variety of configurations and storage devices. The clients and servers were connected with 10 GbE which provided sufficient bandwidth and throughput for reasonably high CPU utilization. The large sort workload gets very little benefit from the local file cache, so its performance did not change with either the SAS FILELOCK default setting or with FILELOCKS=NONE. The calibration workload gets a large benefit from the local file cache and the overall elapsed time for the job increased by 25% or more with the SAS default setting for file locks compared to running with FILELOCKS=NONE. Please note that file locks were only requested for permanent data and that SAS WORK files were accessed with no file locks. For SAS programs which do not rely on file system locks to prevent multiple concurrent read/write access to a file, we recommend FILELOCKS=NONE to improve NFS s local file cache behavior. With FILELOCKS=NONE, file cache retention is improved. The benchmarks were run on a variety of devices including EMC Isilon and Oracle Exalogic. In general, there are few tuning options for NFS, but NFS benefits from isolating file system metadata from file system data. Some storage devices like Isilon allow this isolation. Within the continuum of NFS performance - better results were obtained from devices that both isolate file system metadata and utilize SSD's for file system metadata. Some NFS devices generated errors under the load generated by the calibration workload. SAS R&D is working with several vendors to resolve these issues. SAS technical support is aware of these issues and can be contacted to determine if there are any known issues with any particular device. NFS can be a reasonable choice for small SAS Grid Manager implementations and / or when performance is less of a concern for the customer. NFS appliances have differing architectures and therefore differing performance characteristics. In general it is not a good idea to place SASWORK on an NFS file system where performance is a concern. If customers are concerned about performance they are strongly advised to consider more performant options such as other shared file system choices outlined in this paper. If not taking advantage of checkpoint / restart functionality using NFS for SASDATA and local SSDs for example for SASWORK can provide performance advantages however such an arrangement does come with the tradeoff of needing to manage the mix of jobs and the concurrent demand for file space on the local storage. Global File System2 (Red Hat Resilient Storage) Global File System 2 (GFS2) is a cluster file system supported by Red Hat. GFS2 is provided in the RHEL Resilient Storage Add-On. Note that Red Hat strongly recommends an architecture review be conducted by Red Hat prior to the implementation of a RHEL Cluster. At the time of this update, the cost of the architecture review is covered by the GFS2 support subscription with Red Hat. To initiate an architecture review open a support ticket with Red Hat requesting an architecture review for Resilient Storage. A previous release of this paper indicated performance concerns with GFS2, however improvements delivered with RHEL errata patches through mid-march 2013 have provided marked enhancements to the performance of GFS2. If SAS WORK directories need to be shared then the recommended configuration is to place those directories on a different GFS2 file system distinct from the GFS2 file system used for shared permanent files. GFS2 is limited to 16 nodes in a cluster.

13 A Survey of Shared File Systems 9 The workloads were run using the tuned utility with the enterprise-storage profile. This configures the system to use the deadline I/O scheduler and sets the dirty page ratio to 40 by setting the option VM.DIRTY_RATIO= to 40. These settings greatly improved the performance of the workloads. To improve the behavior of the Distributed Lock Manager (DLM), the following parameters were set to 16384: DLM_LKBTBL_SIZE DLM_RSBTBL_SIZE DLM_DIRTBL_SIZE Change these values by editing /etc/init.d/cman to ensure that the values are retained upon each reboot of the system. The benchmark environment used the LVCHANGE-R command <the value should be appropriate for the workload> to set read-ahead for the file system. We recommend setting read ahead for SAS workloads. Other commands may be used to set read ahead and more detail can be found in the references below. For optimal GFS2 performance run RHEL 6.4 and ensure that all patches are applied at least through mid-march These errata provide fixes to the tuned service as well as address a concern with irqbalance. With RHEL 6.4 and the mid- March 2013 errata SAS found that the workload performed very well on GFS2. This version of GFS2 has not displayed rapid performance degradation observed with prior versions. The other shared file systems measured on the RHEL platform were GPFS, StorNext and NFS. GPFS and StorNext are excellent choices when there is a need for extreme performance or the need to have greater than 16 nodes in the cluster. GFS2 performance has improved markedly and is now a very good choice for clusters up to 16 nodes. NFS is an option when performance is not a primary concern. See the following URLs for additional information: Common Internet File System (CIFS) Common Internet File System (CIFS) is the native shared file system provided with Windows operating systems. With recent patches, CIFS can be used for workloads with moderate levels of concurrency and works best for workloads that get limited benefit from the local file cache. The recommended configuration is to place SAS WORK directories on a non- CIFS file system and use CIFS to manage shared permanent files. With the release of the Windows Vista operating system in 2006, many improvements were made to address performance issues and connectivity via 10 GbE greatly improves the throughput and responsiveness of the CIFS file system. With changes made both to SAS Foundation 9.3 software and Windows Server 2008 R2 operating system, CIFS is functionally stable. However, benchmark results showed there was relatively poor retention of data in the local file cache. Workloads that reuse data from local file cache will not perform nearly as well with CIFS compared to a local file system. The benchmark configuration had three systems running the Windows Server 2008 R2 operating system, one acting as file server and two systems as clients all connected via 10 GbE. In order to function properly and perform optimally, the following registry settings, Microsoft patches, and environment variable settings used by SAS 9.3 software should be used:

14 10 A Survey of Shared File Systems [HKLM\SYSTEM\CurrentControlSet\Control\SessionManager\Executive] AdditionalDelayedWorkerThreads = 0x28 AdditionalCriticalWorkerThreads = 0x28 [HKLM\SYSTEM\CurrentControlSet\Services\lanmanworkstation\parameters] DirectoryCacheLifetime=0x0 FileNotFoundCacheLifetime=0x0 Hotfix referenced by Hotfix referenced by Environment variable SAS_NO_RANDOM_ACCESS = 1 Hotfix disabled file read-ahead under some circumstances. Microsoft will release an additional patch in April, 2012 to address this issue. Comparing CIFS to other shared file systems measured on the Windows platform, GPFS had much better throughput, local file cache retention and metadata management compared to CIFS. This translated to better scalability and GPFS can serve a larger number of client systems compared to CIFS. Quantum StorNext StorNext is a file system and tiered data management family of products provided by Quantum Corporation. It is supported on a variety of operating systems including Red Hat Enterprise Linux and Microsoft Windows operating systems. Benchmarks were run with both Quantum StorNext 4.3 and a pre-release copy of StorNext 5 on Red Hat Enterprise Linux (RHEL) 6.4. Both versions performed well, but StorNext 5 contains additional performance and scalability improvements as well as updates that address storage capacity issues encountered with heavy file metadata environments in the earlier release. (StorNext 5 is scheduled to be available in StorNext appliances in CY2013. Availability of the software only solution is targeted for early 2014) The configuration utilized separate StorNext file systems for permanent data and SAS WORK files with excellent throughput. The demand model forecasting and calibration workload creates and deletes a high number of files in a relatively short period of time. The 4.3 version required more temporary storage capacity because artifacts from deleted files were not completely eliminated from the system as quickly as new files were created. StorNext 5 pre-release software solved this issue by significantly improving file creation and deletion performance as well as achieving low overhead for file system metadata management. The configuration utilized 8 systems running the SAS workloads, with 2 of these systems also acting as StorNext metadata controllers. Both file systems had a blocksize of 16k. The file system for permanent data had a journal size of 64M and the SAS WORK file system had a journal size of 512M. The following options were set in the cfgx file for each of the two file systems: filelocks true (default false) globalsuperuser true (default false)

15 A Survey of Shared File Systems 11 maxconnections 64 (default 10) These options were increased from default values to improve performance: inodecachesize 512k (default 32k) buffercachesize 4G (default 32M) threadpoolsize 1024 (default 16) The SAS WORK file system also used ASR (Allocation Session Reservation) set to 10G. The default is 0 which is off. The following mount options were changed from default values: nodiratime and noatime are set, default is not set. threads 32 (default 12) cachebufsize 128k (default 32k) buffercachecap 30730M (default varies by hardware, would be 256M for a 64 bit system with > 2G of memory ) buffercache_readahead 24 (default 16) buffercache_iods 32 (default 8) dircachesize 64M (default 10M) auto_dma_read_length and auto_dma_write_length 2048 but testing defaulted to a minimum value of cachebufsize+1 (128k+1) Overall, the StorNext file systems behaved very well, with excellent usage of the buffer cache and low overhead for file system metadata operations. Intel Enterprise Edition for Lustre Intel Enterprise Edition for Lustre Version 1.0 (IEEL V1.0) is a commercial version of the parallel file system Lustre supported by Intel. Earlier benchmarks run with other versions of Lustre encountered several performance problems, but IEEL V1.0 performed well, with excellent throughput and low overhead for file system metadata management. The benchmark configuration utilized one Metadata Server (MDS), three Object Storage Servers (OSS) and 12 RHEL 6.4 client systems. There was 30 Gb of network bandwidth available between the servers and clients. The Object Storage Targets were constructed from a disk array with k rpm disks. The Metadata Target was constructed from 4 SSD s configured as RAID 10. Intel recommends 128 GB of memory for each server, however the benchmark configuration had only 24 GB for each server. The demand model forecasting and calibration workload generates a high amount of metadata requests. For each mount point on a client system, there is a single metadata request queue. This was found to be a bottleneck for the initial configuration. A single Lustre file system can be mounted multiple times. To create additional metadata queues, the file system was mounted four times and three of the mount points were used for SAS work directories. The SAS configuration file was modified so that the work option was set to a file that contained a list of three directories and the method for

16 12 A Survey of Shared File Systems selecting a work directory was set to random. The additional metadata queues eliminated the metadata bottleneck. The file system was formatted using default options. Several file system settings were modified for best performance of the benchmarks. On the OSS the option readcache_max_filesize controls the maximum size of a file that both the read cache and writethrough cache will try to keep in memory. Files larger than readcache_max_filesize will not be kept in cache for either reads or writes. This was set to 2M on each OSS via the command: lctl set_param obdfilter.*.readcache_max_filesize=2m On the client systems, the maximum for dirty data that can be written and queued was increased from the default to 256 MB. Also, the maximum number of in-flight RPC s was increased from the default to 16. These changes were made using the following commands: lctl set_param osc.\*.max_dirty_mb=256 lctl set_param osc.\*.max_rpcs_in_flight=16 Overall, the Intel Enterprise Edition for Lustre Version 1.0 file systems behaved very well, with excellent usage of the local file system cache and low overhead for file system metadata operations. Veritas Cluster File System The SAS Performance Analysis lab has not done first hand experiments with Veritas products (owned and maintained by Symantec since 2005) however we have experience with this file system in the field via customer engagements. The following are best practices documented by SAS employees working with customer implementations of SAS and the Veritas file system in the field: 1. Use 8KB block size when creating a VxFS file system 2. Use multipathing. 3. Separate permanent SAS data and temporary SASWORK data on different file systems. 4. When creating a Veritas volume use a RAID-0 stripe across multiple LUNs (assuming LUNS are RAID protected). 5. Depending upon the underlying LUN characteristics make the VxVM strip size be equal to a full stripe-width on the underlying LUN. For example if a LUN is a RAID-5 volume created as a 4+1 with a 64KB strip size, the strip size for the host should be 256KB. 6. In the above case the SAS BUFSIZE parameter should be set to 256KB in the SAS configuration file. 7. If a default SAS BUFSIZE is 256KB or larger use vxtunefs to increase the value of discovery_direct_iosz. 8. If SAS BUFSIZE is 256KB or less change the vxtunefs parameters Read_perf_io Write_perf_io to match the SAS BUFSIZE otherwise make the values an integer divisor of the SAS BUFSIZE. 9. Check read_nstream and write_nstream values are set appropriately from VxVM volume configuration. 10. Mount the SASWORK temporary file system(s) with tmplog and mincache=tmpcache. In SAS Version 9.1 and 9.2 have users modify LIBNAME statement to add offset=value, where value is the SAS BUFSIZE. Using the example above (5) add offset=256k to the LIBNAME statement. In version 9.3 and above put alignsasiofiles in the sasv9 config file. By default a SAS data set contains 512 bytes of header data. The actual SAS data begins in the file at an 8K offset and

17 A Survey of Shared File Systems 13 the file is written to with BUFSIZE IO request sizes. When offset= is set then the first data page will be at offset in the file. For very large files that are accessed by a single SAS job at any point in time consider using direction as documented in Conclusion This paper contains only the file systems that we have tuned and measured in our performance lab. Other shared file systems have been used successfully in distributed deployments, such as Veritas Cluster File System. We plan to continue running benchmarks against other configurations and will publish the results as they become available. Resources Windows is a registered trademark of Microsoft Corporation in the United States and other countries. Windows Vista is either a registered trademark or trademark of Microsoft Corporation in the United States and/or other countries. DDN, SFA, 10K-E, and In-Storage Processing are trademarks owned by DataDirect Networks.

18 14 A Survey of Shared File Systems SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks of SAS Institute Inc. in the USA and other countries. indicates USA registration. Other brand and product names are trademarks of their respective companies. Copyright 2010 SAS Institute Inc., Cary, NC, USA. All rights reserved.

A Survey of Shared File Systems

A Survey of Shared File Systems Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications A Survey of Shared File Systems A Survey of Shared File Systems Table of Contents Introduction...

More information

HP reference configuration for entry-level SAS Grid Manager solutions

HP reference configuration for entry-level SAS Grid Manager solutions HP reference configuration for entry-level SAS Grid Manager solutions Up to 864 simultaneous SAS jobs and more than 3 GB/s I/O throughput Technical white paper Table of contents Executive summary... 2

More information

Architecting Your SAS Grid : Networking for Performance

Architecting Your SAS Grid : Networking for Performance ABSTRACT Paper SAS6760-2016 Architecting Your SAS : Networking for Performance Tony Brown and Margaret Crevar, SAS Institute Inc. With the popularity of network-attached storage for shared file systems,

More information

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010

Best Practices for Data Sharing in a Grid Distributed SAS Environment. Updated July 2010 Best Practices for Data Sharing in a Grid Distributed SAS Environment Updated July 2010 B E S T P R A C T I C E D O C U M E N T Table of Contents 1 Abstract... 2 1.1 Storage performance is critical...

More information

Integrated Grid Solutions. and Greenplum

Integrated Grid Solutions. and Greenplum EMC Perspective Integrated Grid Solutions from SAS, EMC Isilon and Greenplum Introduction Intensifying competitive pressure and vast growth in the capabilities of analytic computing platforms are driving

More information

How to Choose your Red Hat Enterprise Linux Filesystem

How to Choose your Red Hat Enterprise Linux Filesystem How to Choose your Red Hat Enterprise Linux Filesystem EXECUTIVE SUMMARY Choosing the Red Hat Enterprise Linux filesystem that is appropriate for your application is often a non-trivial decision due to

More information

EMC Unified Storage for Microsoft SQL Server 2008

EMC Unified Storage for Microsoft SQL Server 2008 EMC Unified Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON and EMC FAST Cache Reference Copyright 2010 EMC Corporation. All rights reserved. Published October, 2010 EMC believes the information

More information

SAS System for Windows: Integrating with a Network Appliance Filer Mark Hayakawa, Technical Marketing Engineer, Network Appliance, Inc.

SAS System for Windows: Integrating with a Network Appliance Filer Mark Hayakawa, Technical Marketing Engineer, Network Appliance, Inc. Paper 220-29 SAS System for Windows: Integrating with a Network Appliance Filer Mark Hayakawa, Technical Marketing Engineer, Network Appliance, Inc. ABSTRACT The goal of this paper is to help customers

More information

Scale and Availability Considerations for Cluster File Systems. David Noy, Symantec Corporation

Scale and Availability Considerations for Cluster File Systems. David Noy, Symantec Corporation Scale and Availability Considerations for Cluster File Systems David Noy, Symantec Corporation SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted.

More information

Best Practices for Configuring Your I/O Subsystem for SAS 9 Applications

Best Practices for Configuring Your I/O Subsystem for SAS 9 Applications Paper SAS1501-2015 Best Practices for Configuring Your I/O Subsystem for SAS 9 Applications Tony Brown, Margaret Crevar, SAS Institute Inc., Cary, NC ABSTRACT The power of SAS 9 applications allows information

More information

SAS Analytics on IBM FlashSystem storage: Deployment scenarios and best practices

SAS Analytics on IBM FlashSystem storage: Deployment scenarios and best practices Paper 3290-2015 SAS Analytics on IBM FlashSystem storage: Deployment scenarios and best practices ABSTRACT Harry Seifert, IBM Corporation; Matt Key, IBM Corporation; Narayana Pattipati, IBM Corporation;

More information

Technical Paper. Performance and Tuning Considerations for SAS on the EMC XtremIO TM All-Flash Array

Technical Paper. Performance and Tuning Considerations for SAS on the EMC XtremIO TM All-Flash Array Technical Paper Performance and Tuning Considerations for SAS on the EMC XtremIO TM All-Flash Array Release Information Content Version: 1.0 October 2014. Trademarks and Patents SAS Institute Inc., SAS

More information

Benchmarking Hadoop & HBase on Violin

Benchmarking Hadoop & HBase on Violin Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages

More information

SAS Grid Manager Testing and Benchmarking Best Practices for SAS Intelligence Platform

SAS Grid Manager Testing and Benchmarking Best Practices for SAS Intelligence Platform SAS Grid Manager Testing and Benchmarking Best Practices for SAS Intelligence Platform INTRODUCTION Grid computing offers optimization of applications that analyze enormous amounts of data as well as load

More information

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

HP SN1000E 16 Gb Fibre Channel HBA Evaluation HP SN1000E 16 Gb Fibre Channel HBA Evaluation Evaluation report prepared under contract with Emulex Executive Summary The computing industry is experiencing an increasing demand for storage performance

More information

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000

The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000 The IntelliMagic White Paper: Storage Performance Analysis for an IBM Storwize V7000 Summary: This document describes how to analyze performance on an IBM Storwize V7000. IntelliMagic 2012 Page 1 This

More information

recovery at a fraction of the cost of Oracle RAC

recovery at a fraction of the cost of Oracle RAC Concurrent data access and fast failover for unstructured data and Oracle databases recovery at a fraction of the cost of Oracle RAC Improve application performance and scalability - Get parallel processing

More information

CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY

CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY White Paper CONFIGURATION GUIDELINES: EMC STORAGE FOR PHYSICAL SECURITY DVTel Latitude NVMS performance using EMC Isilon storage arrays Correct sizing for storage in a DVTel Latitude physical security

More information

Object storage in Cloud Computing and Embedded Processing

Object storage in Cloud Computing and Embedded Processing Object storage in Cloud Computing and Embedded Processing Jan Jitze Krol Systems Engineer DDN We Accelerate Information Insight DDN is a Leader in Massively Scalable Platforms and Solutions for Big Data

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

Big data management with IBM General Parallel File System

Big data management with IBM General Parallel File System Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers

More information

Performance and scalability of a large OLTP workload

Performance and scalability of a large OLTP workload Performance and scalability of a large OLTP workload ii Performance and scalability of a large OLTP workload Contents Performance and scalability of a large OLTP workload with DB2 9 for System z on Linux..............

More information

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Using Synology SSD Technology to Enhance System Performance Synology Inc. Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...

More information

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms

Distributed File System. MCSN N. Tonellotto Complements of Distributed Enabling Platforms Distributed File System 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributed File System Don t move data to workers move workers to the data! Store data on the local disks of nodes

More information

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs Distributed File System Choices: Red Hat Storage, GFS2 & pnfs Ric Wheeler Architect & Senior Manager, Red Hat June 27, 2012 Overview Distributed file system basics Red Hat distributed file systems Performance

More information

Technical Paper. Performance and Tuning Considerations for SAS on Pure Storage FA-420 Flash Array

Technical Paper. Performance and Tuning Considerations for SAS on Pure Storage FA-420 Flash Array Technical Paper Performance and Tuning Considerations for SAS on Pure Storage FA-420 Flash Array Release Information Content Version: 1.0 August 2014. Trademarks and Patents SAS Institute Inc., SAS Campus

More information

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator

More information

Running a Workflow on a PowerCenter Grid

Running a Workflow on a PowerCenter Grid Running a Workflow on a PowerCenter Grid 2010-2014 Informatica Corporation. No part of this document may be reproduced or transmitted in any form, by any means (electronic, photocopying, recording or otherwise)

More information

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director brett.weninger@adurant.com Dave Smelker, Managing Principal dave.smelker@adurant.com

More information

Quantum StorNext. Product Brief: Distributed LAN Client

Quantum StorNext. Product Brief: Distributed LAN Client Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without

More information

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Oracle Database Scalability in VMware ESX VMware ESX 3.5 Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises

More information

High Performance Computing OpenStack Options. September 22, 2015

High Performance Computing OpenStack Options. September 22, 2015 High Performance Computing OpenStack PRESENTATION TITLE GOES HERE Options September 22, 2015 Today s Presenters Glyn Bowden, SNIA Cloud Storage Initiative Board HP Helion Professional Services Alex McDonald,

More information

Technical Paper. Best Practices for SAS on EMC SYMMETRIX VMAX TM Storage

Technical Paper. Best Practices for SAS on EMC SYMMETRIX VMAX TM Storage Technical Paper Best Practices for SAS on EMC SYMMETRIX VMAX TM Storage Paper Title Table of Contents Introduction... 1 BRIEF OVERVIEW OF VMAX ARCHITECTURE... 1 PHYSICAL STORAGE DISK TYPES, FA PORTS,

More information

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute PADS GPFS Filesystem: Crash Root Cause Analysis Computation Institute Argonne National Laboratory Table of Contents Purpose 1 Terminology 2 Infrastructure 4 Timeline of Events 5 Background 5 Corruption

More information

Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card

Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card Version 1.0 April 2011 DB15-000761-00 Revision History Version and Date Version 1.0, April 2011 Initial

More information

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads 89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com @EdisonGroupInc 212.367.7400 IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads A Competitive Test and Evaluation Report

More information

HP Smart Array Controllers and basic RAID performance factors

HP Smart Array Controllers and basic RAID performance factors Technical white paper HP Smart Array Controllers and basic RAID performance factors Technology brief Table of contents Abstract 2 Benefits of drive arrays 2 Factors that affect performance 2 HP Smart Array

More information

Software-defined Storage Architecture for Analytics Computing

Software-defined Storage Architecture for Analytics Computing Software-defined Storage Architecture for Analytics Computing Arati Joshi Performance Engineering Colin Eldridge File System Engineering Carlos Carrero Product Management June 2015 Reference Architecture

More information

Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database

Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database WHITE PAPER Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Executive

More information

A Shared File System on SAS Grid Manger in a Cloud Environment

A Shared File System on SAS Grid Manger in a Cloud Environment A Shared File System on SAS Grid Manger in a Cloud Environment ABSTRACT Ying-ping (Marie) Zhang, Intel Corporation, Chandler, Arizona Margaret Carver, SAS Institute, Cary, North Carolina Gabriele Paciucci,

More information

The IntelliMagic White Paper on: Storage Performance Analysis for an IBM San Volume Controller (SVC) (IBM V7000)

The IntelliMagic White Paper on: Storage Performance Analysis for an IBM San Volume Controller (SVC) (IBM V7000) The IntelliMagic White Paper on: Storage Performance Analysis for an IBM San Volume Controller (SVC) (IBM V7000) IntelliMagic, Inc. 558 Silicon Drive Ste 101 Southlake, Texas 76092 USA Tel: 214-432-7920

More information

Application Performance for High Performance Computing Environments

Application Performance for High Performance Computing Environments Application Performance for High Performance Computing Environments Leveraging the strengths of Computationally intensive applications With high performance scale out file serving In data storage modules

More information

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1 Performance Study Performance Characteristics of and RDM VMware ESX Server 3.0.1 VMware ESX Server offers three choices for managing disk access in a virtual machine VMware Virtual Machine File System

More information

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies

More information

WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE

WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE 1 W W W. F U S I ON I O.COM Table of Contents Table of Contents... 2 Executive Summary... 3 Introduction: In-Memory Meets iomemory... 4 What

More information

WHITE PAPER FUJITSU PRIMERGY SERVER BASICS OF DISK I/O PERFORMANCE

WHITE PAPER FUJITSU PRIMERGY SERVER BASICS OF DISK I/O PERFORMANCE WHITE PAPER BASICS OF DISK I/O PERFORMANCE WHITE PAPER FUJITSU PRIMERGY SERVER BASICS OF DISK I/O PERFORMANCE This technical documentation is aimed at the persons responsible for the disk I/O performance

More information

Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering

Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations. Database Solutions Engineering Comprehending the Tradeoffs between Deploying Oracle Database on RAID 5 and RAID 10 Storage Configurations A Dell Technical White Paper Database Solutions Engineering By Sudhansu Sekhar and Raghunatha

More information

Windows 8 SMB 2.2 File Sharing Performance

Windows 8 SMB 2.2 File Sharing Performance Windows 8 SMB 2.2 File Sharing Performance Abstract This paper provides a preliminary analysis of the performance capabilities of the Server Message Block (SMB) 2.2 file sharing protocol with 10 gigabit

More information

Introduction to Gluster. Versions 3.0.x

Introduction to Gluster. Versions 3.0.x Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster

More information

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC

More information

Evaluation of Enterprise Data Protection using SEP Software

Evaluation of Enterprise Data Protection using SEP Software Test Validation Test Validation - SEP sesam Enterprise Backup Software Evaluation of Enterprise Data Protection using SEP Software Author:... Enabling you to make the best technology decisions Backup &

More information

Shared Parallel File System

Shared Parallel File System Shared Parallel File System Fangbin Liu fliu@science.uva.nl System and Network Engineering University of Amsterdam Shared Parallel File System Introduction of the project The PVFS2 parallel file system

More information

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup

More information

VIDEO SURVEILLANCE WITH SURVEILLUS VMS AND EMC ISILON STORAGE ARRAYS

VIDEO SURVEILLANCE WITH SURVEILLUS VMS AND EMC ISILON STORAGE ARRAYS VIDEO SURVEILLANCE WITH SURVEILLUS VMS AND EMC ISILON STORAGE ARRAYS Successfully configure all solution components Use VMS at the required bandwidth for NAS storage Meet the bandwidth demands of a 2,200

More information

AIX NFS Client Performance Improvements for Databases on NAS

AIX NFS Client Performance Improvements for Databases on NAS AIX NFS Client Performance Improvements for Databases on NAS October 20, 2005 Sanjay Gulabani Sr. Performance Engineer Network Appliance, Inc. gulabani@netapp.com Diane Flemming Advisory Software Engineer

More information

Technical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment

Technical Paper. Moving SAS Applications from a Physical to a Virtual VMware Environment Technical Paper Moving SAS Applications from a Physical to a Virtual VMware Environment Release Information Content Version: April 2015. Trademarks and Patents SAS Institute Inc., SAS Campus Drive, Cary,

More information

Intel Data Direct I/O Technology (Intel DDIO): A Primer >

Intel Data Direct I/O Technology (Intel DDIO): A Primer > Intel Data Direct I/O Technology (Intel DDIO): A Primer > Technical Brief February 2012 Revision 1.0 Legal Statements INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL PRODUCTS. NO LICENSE,

More information

Q & A From Hitachi Data Systems WebTech Presentation:

Q & A From Hitachi Data Systems WebTech Presentation: Q & A From Hitachi Data Systems WebTech Presentation: RAID Concepts 1. Is the chunk size the same for all Hitachi Data Systems storage systems, i.e., Adaptable Modular Systems, Network Storage Controller,

More information

Configuring RAID for Optimal Performance

Configuring RAID for Optimal Performance Configuring RAID for Optimal Performance Intel RAID Controller SRCSASJV Intel RAID Controller SRCSASRB Intel RAID Controller SRCSASBB8I Intel RAID Controller SRCSASLS4I Intel RAID Controller SRCSATAWB

More information

A Shared File System on SAS Grid Manager * in a Cloud Environment

A Shared File System on SAS Grid Manager * in a Cloud Environment white paper A Shared File System on SAS Grid Manager * in a Cloud Environment Abstract The performance of a shared file system is a critical component of implementing SAS Grid Manager* in a cloud environment.

More information

WHITE PAPER Optimizing Virtual Platform Disk Performance

WHITE PAPER Optimizing Virtual Platform Disk Performance WHITE PAPER Optimizing Virtual Platform Disk Performance Think Faster. Visit us at Condusiv.com Optimizing Virtual Platform Disk Performance 1 The intensified demand for IT network efficiency and lower

More information

Using Synology SSD Technology to Enhance System Performance. Based on DSM 5.2

Using Synology SSD Technology to Enhance System Performance. Based on DSM 5.2 Using Synology SSD Technology to Enhance System Performance Based on DSM 5.2 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD Cache as Solution...

More information

Accelerating Server Storage Performance on Lenovo ThinkServer

Accelerating Server Storage Performance on Lenovo ThinkServer Accelerating Server Storage Performance on Lenovo ThinkServer Lenovo Enterprise Product Group April 214 Copyright Lenovo 214 LENOVO PROVIDES THIS PUBLICATION AS IS WITHOUT WARRANTY OF ANY KIND, EITHER

More information

Netapp @ 10th TF-Storage Meeting

Netapp @ 10th TF-Storage Meeting Netapp @ 10th TF-Storage Meeting Wojciech Janusz, Netapp Poland Bogusz Błaszkiewicz, Netapp Poland Ljubljana, 2012.02.20 Agenda Data Ontap Cluster-Mode pnfs E-Series NetApp Confidential - Internal Use

More information

Amadeus SAS Specialists Prove Fusion iomemory a Superior Analysis Accelerator

Amadeus SAS Specialists Prove Fusion iomemory a Superior Analysis Accelerator WHITE PAPER Amadeus SAS Specialists Prove Fusion iomemory a Superior Analysis Accelerator 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com SAS 9 Preferred Implementation Partner tests a single Fusion

More information

Virtuoso and Database Scalability

Virtuoso and Database Scalability Virtuoso and Database Scalability By Orri Erling Table of Contents Abstract Metrics Results Transaction Throughput Initializing 40 warehouses Serial Read Test Conditions Analysis Working Set Effect of

More information

COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters

COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters COSC 6374 Parallel I/O (I) I/O basics Fall 2012 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network card 1 Network card

More information

An Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine

An Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine An Oracle White Paper May 2011 Exadata Smart Flash Cache and the Oracle Exadata Database Machine Exadata Smart Flash Cache... 2 Oracle Database 11g: The First Flash Optimized Database... 2 Exadata Smart

More information

Optimizing LTO Backup Performance

Optimizing LTO Backup Performance Optimizing LTO Backup Performance July 19, 2011 Written by: Ash McCarty Contributors: Cedrick Burton Bob Dawson Vang Nguyen Richard Snook Table of Contents 1.0 Introduction... 3 2.0 Host System Configuration...

More information

Configuring EMC CLARiiON for SAS Business Intelligence Platforms

Configuring EMC CLARiiON for SAS Business Intelligence Platforms Configuring EMC CLARiiON for SAS Business Intelligence Platforms Applied Technology Abstract Deploying SAS applications optimally with data stored on EMC CLARiiON systems requires a close collaboration

More information

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup

Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Technical white paper Using HP StoreOnce Backup Systems for NDMP backups with Symantec NetBackup Table of contents Executive summary... 2 Introduction... 2 What is NDMP?... 2 Technology overview... 3 HP

More information

High Performance SQL Server with Storage Center 6.4 All Flash Array

High Performance SQL Server with Storage Center 6.4 All Flash Array High Performance SQL Server with Storage Center 6.4 All Flash Array Dell Storage November 2013 A Dell Compellent Technical White Paper Revisions Date November 2013 Description Initial release THIS WHITE

More information

Systems Architecture. Paper 271-26 ABSTRACT

Systems Architecture. Paper 271-26 ABSTRACT Paper 271-26 Findings and Sizing Considerations of an Integrated BI Software Solution in an AIX Environment and a Windows NT Environment E. Hayes-Hall, IBM Corporation, Austin, TX ABSTRACT This paper presents

More information

Taking Linux File and Storage Systems into the Future. Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated

Taking Linux File and Storage Systems into the Future. Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated Taking Linux File and Storage Systems into the Future Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated 1 Overview Going Bigger Going Faster Support for New Hardware Current Areas

More information

Technology Insight Series

Technology Insight Series Evaluating Storage Technologies for Virtual Server Environments Russ Fellows June, 2010 Technology Insight Series Evaluator Group Copyright 2010 Evaluator Group, Inc. All rights reserved Executive Summary

More information

HPSA Agent Characterization

HPSA Agent Characterization HPSA Agent Characterization Product HP Server Automation (SA) Functional Area Managed Server Agent Release 9.0 Page 1 HPSA Agent Characterization Quick Links High-Level Agent Characterization Summary...

More information

Storage Management for the Oracle Database on Red Hat Enterprise Linux 6: Using ASM With or Without ASMLib

Storage Management for the Oracle Database on Red Hat Enterprise Linux 6: Using ASM With or Without ASMLib Storage Management for the Oracle Database on Red Hat Enterprise Linux 6: Using ASM With or Without ASMLib Sayan Saha, Sue Denham & Lars Herrmann 05/02/2011 On 22 March 2011, Oracle posted the following

More information

Protect Microsoft Exchange databases, achieve long-term data retention

Protect Microsoft Exchange databases, achieve long-term data retention Technical white paper Protect Microsoft Exchange databases, achieve long-term data retention HP StoreOnce Backup systems, HP StoreOnce Catalyst, and Symantec NetBackup OpenStorage Table of contents Introduction...

More information

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software The Video Edition of XenData Archive Series software manages one or more automated data tape libraries on

More information

Partition Alignment Dramatically Increases System Performance

Partition Alignment Dramatically Increases System Performance Partition Alignment Dramatically Increases System Performance Information for anyone in IT that manages large storage environments, data centers or virtual servers. Paragon Software Group Paragon Alignment

More information

Cray DVS: Data Virtualization Service

Cray DVS: Data Virtualization Service Cray : Data Virtualization Service Stephen Sugiyama and David Wallace, Cray Inc. ABSTRACT: Cray, the Cray Data Virtualization Service, is a new capability being added to the XT software environment with

More information

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE White Paper IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE Abstract This white paper focuses on recovery of an IBM Tivoli Storage Manager (TSM) server and explores

More information

Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010

Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010 Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010 This document is provided as-is. Information and views expressed in this document, including URL and other Internet

More information

GiantLoop Testing and Certification (GTAC) Lab

GiantLoop Testing and Certification (GTAC) Lab GiantLoop Testing and Certification (GTAC) Lab Benchmark Test Results: VERITAS Foundation Suite and VERITAS Database Edition Prepared For: November 2002 Project Lead: Mike Schwarm, Director, GiantLoop

More information

Technical Paper. Performance and Tuning Considerations for SAS on Fusion-io ioscale Flash Storage

Technical Paper. Performance and Tuning Considerations for SAS on Fusion-io ioscale Flash Storage Technical Paper Performance and Tuning Considerations for SAS on Fusion-io ioscale Flash Storage Release Information Content Version: 1.0 May 2014. Trademarks and Patents SAS Institute Inc., SAS Campus

More information

EMC VFCACHE ACCELERATES ORACLE

EMC VFCACHE ACCELERATES ORACLE White Paper EMC VFCACHE ACCELERATES ORACLE VFCache extends Flash to the server FAST Suite automates storage placement in the array VNX protects data EMC Solutions Group Abstract This white paper describes

More information

Introduction to NetApp Infinite Volume

Introduction to NetApp Infinite Volume Technical Report Introduction to NetApp Infinite Volume Sandra Moulton, Reena Gupta, NetApp April 2013 TR-4037 Summary This document provides an overview of NetApp Infinite Volume, a new innovation in

More information

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance. Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance

More information

SAS Business Analytics. Base SAS for SAS 9.2

SAS Business Analytics. Base SAS for SAS 9.2 Performance & Scalability of SAS Business Analytics on an NEC Express5800/A1080a (Intel Xeon 7500 series-based Platform) using Red Hat Enterprise Linux 5 SAS Business Analytics Base SAS for SAS 9.2 Red

More information

June 2009. Blade.org 2009 ALL RIGHTS RESERVED

June 2009. Blade.org 2009 ALL RIGHTS RESERVED Contributions for this vendor neutral technology paper have been provided by Blade.org members including NetApp, BLADE Network Technologies, and Double-Take Software. June 2009 Blade.org 2009 ALL RIGHTS

More information

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Using Synology SSD Technology to Enhance System Performance Synology Inc. Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_WP_ 20121112 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD

More information

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle

Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Direct NFS - Design considerations for next-gen NAS appliances optimized for database workloads Akshay Shah Gurmeet Goindi Oracle Agenda Introduction Database Architecture Direct NFS Client NFS Server

More information

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers

Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers Converged storage architecture for Oracle RAC based on NVMe SSDs and standard x86 servers White Paper rev. 2015-11-27 2015 FlashGrid Inc. 1 www.flashgrid.io Abstract Oracle Real Application Clusters (RAC)

More information

HIGHLY AVAILABLE MULTI-DATA CENTER WINDOWS SERVER SOLUTIONS USING EMC VPLEX METRO AND SANBOLIC MELIO 2010

HIGHLY AVAILABLE MULTI-DATA CENTER WINDOWS SERVER SOLUTIONS USING EMC VPLEX METRO AND SANBOLIC MELIO 2010 White Paper HIGHLY AVAILABLE MULTI-DATA CENTER WINDOWS SERVER SOLUTIONS USING EMC VPLEX METRO AND SANBOLIC MELIO 2010 Abstract This white paper demonstrates key functionality demonstrated in a lab environment

More information

Ultimate Guide to Oracle Storage

Ultimate Guide to Oracle Storage Ultimate Guide to Oracle Storage Presented by George Trujillo George.Trujillo@trubix.com George Trujillo Twenty two years IT experience with 19 years Oracle experience. Advanced database solutions such

More information

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution EMC Virtual Infrastructure for Microsoft Applications Data Center Solution Enabled by EMC Symmetrix V-Max and Reference Architecture EMC Global Solutions Copyright and Trademark Information Copyright 2009

More information

Understanding Disk Storage in Tivoli Storage Manager

Understanding Disk Storage in Tivoli Storage Manager Understanding Disk Storage in Tivoli Storage Manager Dave Cannon Tivoli Storage Manager Architect Oxford University TSM Symposium September 2005 Disclaimer Unless otherwise noted, functions and behavior

More information

NetApp High-Performance Computing Solution for Lustre: Solution Guide

NetApp High-Performance Computing Solution for Lustre: Solution Guide Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5

More information

TU04. Best practices for implementing a BI strategy with SAS Mike Vanderlinden, COMSYS IT Partners, Portage, MI

TU04. Best practices for implementing a BI strategy with SAS Mike Vanderlinden, COMSYS IT Partners, Portage, MI TU04 Best practices for implementing a BI strategy with SAS Mike Vanderlinden, COMSYS IT Partners, Portage, MI ABSTRACT Implementing a Business Intelligence strategy can be a daunting and challenging task.

More information