SFA Product Line. ddn.com. DDN Whitepaper. High Performance Solutions for Big Data: Setting the Bar in Both Bandwidth & lops

Size: px
Start display at page:

Download "SFA Product Line. ddn.com. DDN Whitepaper. High Performance Solutions for Big Data: Setting the Bar in Both Bandwidth & lops"

Transcription

1 DDN Whitepaper SFA Product Line High Performance Solutions for Big Data: Setting the Bar in Both Bandwidth & lops

2 Table of Contents Abstract 3 Introduction 3 SFA12KX 5 SFA12KX Storage OS Architecture 6 Active/Active Model 6 Data Protection 8 RAID 8 Hot Spares 9 Battery Backed Write-Back Cache 9 Mirrored Write-Back Cache 9 Mirrored Transaction Journal 10 Metadata Mirrored n-ways 10 Parity Check On Read DirectProtect: 10 Data Integrity Field DirectProtect: 11 Storage System Efficiencies 11 Storage Fusion Xcelerator 11 Partial Disk Rebuild 12 Real-time Adaptive Cache Technology (ReACT ) 12 Rebuild Priority 13 Read Quality of Service 13 Management 14 DirectMon 14 Application Programming Interface (API) 14 Software Summary 14 SFA12KX Hardware Architecture 15 RAID Processing 15 I/O Channels and Architecture 16 Cache 16 Back End Disk Connectivity 16 Hardware Summary 17 SFA OS and In-Storage Processing Computing Systems 17 In-Storage Processing Computing Capability 17 PCIe Device Dedication 18 Virtual Disk Driver 18 Reduction in Equipment, Infrastructure and Complexity 19 SFA12KX Family: Summary 20 2

3 Abstract Big Data dominates the storage landscape. Storage and compute challenges that once fell mostly in the domain of High Performance Computing (HPC) or even Superomputing are an everyday part of businesses that want to make the most of their Big Data. Storage systems that take on these Big Data challenges generally fall into one of two categories: those with high IOPS capability or those with high bandwidth capability. In the world of HPC, where the focus is on massive scale and streaming writes, storage systems with high bandwidth capabilities have been favored. Big Data often is comprised of large files that also benefit from high bandwidth for ingest or write-out activities. Increasing core counts and ever increasing numbers of nodes in HPC has fundamentally changed data I/O patterns and storage system requirements. Big Data s analytic processing challenges require very high levels of IOPS. Traditional storage systems are not capable of both high IOPS and high bandwidth. This paper presents the next step in the evolution of the SFA storage product line. The SFA12KX Family builds on the concepts and success of the SFA10K Family and is uniquely suited to adapt to modern compute environments and the unique data storage challenges they present. The SFA Family performs at the highest levels of both sequential and random IOPS. Additionally we will examine the architecture that enables In-Storage Processing for embedding clustered/parallel file systems directly into the storage resulting in significant reductions in complexity, latency, footprint and cost. Introduction Across the storage industry, the vast majority of block storage systems have been deliberately designed to deliver random access I/O to serve transactional applications. These applications include reservation systems, banking applications, databases, and messaging applications and batch processing jobs. These compute processes use fixed, structured storage formats, which are commonly referred to as structured data. With structured data, information is communicated to/ from storage in small blocks, and accessed in a generally random pattern which requires high Input/Output Operations per Second (IOPS) to deliver sufficient performance. In recent years, the digital content revolution has enabled personal, social and business computing, as well as the ability for predictive simulation, weather forecasting and processing of satellite imagery. This has resulted in an explosion in both size and number of files stored online. Businesses, both online as well as traditional brick and mortar are collecting data at an astonishing rate and are analyzing this data with various methods such as MapReduce which utilize large data sets and require both high bandwidth and high random IOPS capability. We call this Big Data. According to IDC, the Big Data market is expected to grow to $16.9B in 2015 with a compounded annual grow rate (CAGR) of 39.4%. Among the different segments, storage is the fastest growing with 61.4% CAGR. 3

4 $18,000 $16,000 $14,000 $12,000 $10,000 $8,000 $6,000 Services Software Networking Storage Servers $4,000 $2,000 $ Figure 1 Worldwide Big Data Technology and Services Revenue, IDC March 2012 The growth and emergence of Big Data has necessitated change in storage technology to deliver high random IOPS and high bandwidth at the same time. This market opportunity gave rise to storage architectural platforms such as DDN s Silicon Storage Architecture (S2A) which uses specialized hardware to read and write unstructured data at the highest performance levels both for read and write and with no degradation during system correction events such as drive rebuilds. Just as systems optimized for random IOPS do not excel at storing large sequential files, systems optimized for bandwidth are not necessarily class-leading in transactional data patterns. The explosive growth in unstructured data favors storage systems optimized for bandwidth, as the growth in structured data as a percentage of aggregate market demand is slowing year over year. This growth has largely coincided with increasing CPU speeds. As processor frequency approached the upper limits of what is physically possible with silicon based technology, CPU manufactures found a different avenue to increase compute power per CPU: by combining multiple processing cores onto a single CPU socket, resulting in Moore s Law being extended by several years. Recently, the number of processing cores (or simply cores ) per chip in the commodity space has increased to the point that eight cores are common and a higher number of core processors are just around the corner. 4

5 The increase in the number of cores per chip, and the number of threads per core allows multiple processes to run simultaneously, often producing multiple file accesses at once. What the individual running processes view as sequential access, the storage system sees as increasingly random access as data must be read or written in multiple locations on the storage media rather than stepping sequentially through one location. Further, access to hundreds or thousands of files simultaneously, via a single file system namespace or the effect of thousands of threads writing a single file requires substantial POSIX metadata operations that require high-speed random IOPS for optimal response. The need for multi-threaded, simultaneous file access on a massive scale is not a future requirement, it s happening today. Currently, the top supercomputers have over 1.5 million CPU cores in their compute clusters, resulting in potentially hundreds of thousands of simultaneous file writes during checkpoint operations. Leading websites have tens of billions of files stored; accessed at any time with hundreds of thousands of file accesses per second. The continuous increases in processing cores per socket will allow clients to access more and more files simultaneously. This multi-threaded I/O will produce storage access patterns that are increasingly random, requiring high IOPS capability. Thus, a storage system designed to serve large files to multi-core compute environments must now be optimized to support mixed-mode access, offering both high random IOPS and high bandwidth. Seemingly, storage systems can be optimized to serve either high random IOPS or high bandwidth. Conventional wisdom says that systems can excel at one or the other characteristic, but not both. Conventional wisdom also once said a storage system could not write as fast as it reads at peak bandwidth levels, but DDN s Silicon Storage Architecture proved to breakthrough that long-standing belief. Today, a storage system can offer extreme performance in both random IOPS and bandwidth That system utilizes DDN s new Storage Fusion Architecture and is known as the SFA12KX. SFA12KX SFA12KX builds on the revolutionary Storage Fusion Architecture (SFA), first introduced by DDN with SFA10K. SFA is based on a unique combination of highly parallelized software, industry proven data integrity algorithms and high-speed hardware components to produce a storage controller that performs in the extreme range of both bandwidth and IOPS. By marrying a state of the art, multi-threaded data integrity engine to best of breed in processors, interconnects, buses, memory architecture and media technologies, SFA12KX capitalizes on the same advancements in technology as the clients it serves. This strategy ensures that as these technologies evolve and improve, SFA performance will improve along with them. The SFA12KX employs RAID, data integrity and data management software written from the ground up to take advantage of multi-core processors and modern bus architectures. This highly threaded architecture allows performance to linearly scale with advances in underlying hardware. This same architecture allows the SFA12KX to do what no other RAID controller has been able to do to date: to perform in the extreme range of both bandwidth and IOPS. The SFA12KX delivers random IOPS over 1.7 million burst to cache and over 1.4M sustained 4K IOPS to SSDs. Sequential block bandwidth performance is 48GB/s for simultaneous reads and writes. Designed to house the most scalable unstructured file data, the system supports up to up to 1,680 drives of raw storage while enabling a combination of SAS, SATA or SSD drives. 5

6 16 x FDR InfiniBand Host Ports Optimized Drive Support SFA Interface Virtualization SFA Interface Virtualization SATA GB High-Speed Cache 240 Gb/s Cache Link GB High-Speed Cache Leading Capacity and Cost-Optimized Bandwidth Internal SAS Switching Internal SAS Switching 960 Gb/s Internal SAS Storage Management Network SAS Balanced Mix of IOPS Capacity & Bandwidth P Q RAID 5,6 RAID 6 RAID 5, 6 SSD P RAID 5,6 Q RAID 6 SFA RAID 5, 6 Delivering Unrivaled IOPS for Transactional Applications 1 1 m SFA RAID 1 Figure 2 SFA12KX Active/Active RAID Controller Architectural Overview SFA12KX Storage OS Architecture The SFA12KX runs on the market proven and mature SFA OS. SFA OS was purpose-built to fully exploit the power of multi-core processors. A storage controller is made up of many components and the design goal of SFA OS was to get the maximum performance out of every component in the system. Thus, it is not only the RAID engine that is optimized, but also the cache engine, data movers, drivers, schedulers and much more. All of these storage subsystems are highly parallelized and multi-threaded, creating a powerful, scalable software architecture that serves as the basis for high performance, high availability and rich features that will grow over time. Active/Active Model From conception, the SFA12KX had been designed to work in an Active/Active fashion. There are essentially two ways to implement Active/Active operation in a redundant RAID controller: Active/Active with Distributed Locking, or Active/Active with Dynamic Routing. 6

7 Active/Active with Distributed Locking is the method that has been used historically for DDN s S2A products. With this method, each logical unit is on-line to both controllers. Both controllers cache data for the logical unit, both controllers access the physical disks that contain the logical unit directly, and distributed locks are used to guarantee storage register semantics and write atomicity. The locks are communicated across an inter-controller link (ICL). Because the S2A is optimized for bandwidth and has relatively little cache, ICL traffic is low and does not impact performance, however experience has shown that distributed locking slows IOPS performance. This is partly due to the ICL communication latency, but has more to do with the lock and cache lookup times. Thus, in a system destined to perform at extreme IOPS levels, a different method had to be implemented. SFA implements an Active/Active host presentation model with routing-based data access and full cache coherency. The SFA OS provides preference indicators and target port groups for its SCSI target implementation and thus has the notion of a preferred controller and the preferred RAID Processor (RP). In this approach, each logical unit is online to both controllers, but only one controller takes primary ownership for a given logical unit at a given time. The controller that masters the logical unit caches data for the logical unit and accesses the physical disks that contain that logical unit s data. Additionally, the controller that masters the logical unit is the preferred controller for that logical unit, and I/O requests received by the nonpreferred controller are forwarded to the controller that masters the logical unit. This intelligent approach to storage management requires no distributed locking. Instead, I/O requests are forwarded (Figure 3). When mirrored write-back caching is performed, the data must be transferred to both controllers and so there are no additional data transfers. Read data does have to be transferred across the ICL for reads that are not sent to the preferred controller, however these reads benefit from the logical unit s read-ahead cache. When in write-thru mode, write data does have to be transferred across the ICL for writes that are not sent to the preferred controller. Write to Preferred Path Read from Preferred Path Write to Non- Preferred Path Read from Non- Preferred Path Logical Disk Client Logical Disk Client Logical Disk Client Logical Disk Client Preferred Path Preferred Path Logical Disk Master Controller ICL Partner Controller Logical Disk Master Controller ICL Partner Controller Logical Disk Master Controller ICL Partner Controller Logical Disk Master Controller ICL Partner Controller Cache Cache Cache Cache Cache Cache Cache Cache Data Transfer Cache Mirror ICL = Inter-Controller Link Figure 3 Active/Active Routing Depicting IO Scenarios 7

8 There are several advantages to the Active/Active with Routing method. The main advantage of this approach is that no distributed locking is required and that leads to better I/O performance and a very clean failover implementation, which leads to enhanced data integrity. Another advantage is that the caching, both read and write, is more efficient and effective because all of the cache data can be found in a single location. Virtual disk clients need at least one path to each controller to allow failover and thus need a multi-path I/O driver to recognize that the logical units presented by the two controllers for one logical disk represent the same logical disk. It is important that the multi-path I/O driver is able to understand the standard SCSI preference indicators and target port groups. Such drivers are readily available for most major operating systems including Microsoft Windows server products and Linux. Each SFA storage pool (aka: RAID set) has a preferred home attribute that allows specification of which controller and RP should master the logical disks or virtual disks that are realized with that storage pool. Each logical disk has a current home attribute that indicates the controller that is actually mastering the logical unit currently, and this will change dynamically during failover and failback or when the preferred home attribute is changed. The SCSI preference indicators dynamically change to reflect the current home and the MPIO drivers are designed to dynamically adapt to changes in the SCSI preference indicators, so a proper MPIO driver will send most I/O requests to the controller that masters the logical unit. Data Protection RAID The SFA OS RAID stack provides protection against single, physical disk failures with RAID-1 or RAID-5 data protection as well as double Physical Disk failures through the use of high-speed RAID-6 protection. Both the SFA RAID 5 and 6 parity protection implementations use a rotating parity scheme. The RAID-5 implementation adds a parity chunk to every stripe using XOR. The RAID-6 implementation adds a P and Q chunk to every stripe where P and Q are calculated with Galois Field arithmetic. Particular attention has been paid to closing all of the write holes 1, the method for doing so goes beyond the scope of this paper. A RAID set is implemented using an integral number of equal sized members, which are whole physical disks. The total number of RAID set members must be the number of RAID set data members plus parity members. A chunk is one or more sequential data blocks from a single RAID set member. Each member is made up of a sequence of chunks. A stripe consists of a set of chunks, the same ordinal chunk from each RAID set member. For RAID 6, two of the stripe s chunks are used for parity ( P and Q ) while the remaining chunks are used for logical disk data. The data and parity members are laid out as shown in Figure 4 to provide load balancing for both reads and writes. This is sometimes referred to as left symmetric. For normal reads, only the data members need to be read. Optionally, one parity disk ( P ) is read and the parity is checked, part of a feature called DirectProtect which guards against silent data corruption which is within the realm of possibility and sporadically witnessed with SATA disk drive technology. 1 1 For RAID 5 and RAID 6, in the event of a system failure while there are active writes, the parity of a stripe may become inconsistent with the data. If this is not detected and repaired before a disk or block fails, data loss may ensue as incorrect parity will be used to reconstruct the missing block in that stripe. This potential vulnerability is sometimes known as the write hole. Battery-backed cache and similar techniques are commonly used to reduce the window of opportunity for this to occur. SGIS 8

9 Chunk Stripe P for 0-7 Q for P for 8-15 Q for P for Q for Additional stripes that maximize the chunks on each member. Member Figure 4 Example of RAID-6 RAID set Layout Hot Spares The SFA OS provides pools of spare physical disks that can be automatically used to replace failed physical disks. By replacing a failed RAID set member automatically, the mean-time to-repair for the RAID set is minimized resulting in improved data reliability. Battery Backed Write-Back Cache SFA OS provides a write-back cache feature that is used to improve I/O performance. Write-back cache data which has not been written to disk is preserved by maintaining power to the cache memory in the event of an AC mains failure long enough to copy the contents of the cache to stable storage. In addition, SFA OS is designed to tolerate a simultaneous AC mains failure and RAID software failure. Mirrored Write-Back Cache SFA OS provides the ability to mirror all write-back cache data such that the failure of a single controller will not result in data loss. A storage administrator can optionally turn off write-back cache mirroring for a RAID set (for higher performance) however data protection is reduced for logical units within that RAID set. 9

10 Mirrored Transaction Journal RAID write holes are prevented by executing stripe updates as ACID (Atomicity, Consistency, Isolation and Durability) 2 transactions so that when they are interrupted by a power failure then they can be recovered from the transaction journal when power is restored. This journal is stored within the write-back cache and thus is mirrored so that when a simultaneous power failure and controller hardware failure occurs, the surviving controller can recover the transactions. Metadata Mirrored n-ways SFA OS stores a copy of storage system metadata on 18 physical disks to minimize the likelihood that its metadata is lost or corrupted. DirectProtect Silent Data Corruption Detection and Avoidance DirectProtect is a trademarked name for techniques that detect and correct data errors made by physical disks. It s particularly valuable when using lower-cost spinning disk, such as SATA drives, which are designed with a lower bit-error-rate requirement than enterprise quality SAS disks. In SFA OS there are two levels of DirectProtect: Parity Check on Read (PCOR) and Data Integrity Field (DIF) Parity Check On Read DirectProtect: The SFA OS allows the administrator to specify whether DirectProtect is turned on or off per RAID set. If enabled for a given RAID set, RAID parity will be checked on all reads. In the event that the RAID parity is found to be bad, SFA OS takes steps to correct the data including retrying the reads and using P and Q to identify the bad data. Once the bad data is identified the correct data is generated from parity and the read is returned. Any bad data on physical disk is corrected in the process. When data is read as part of a write operation (e.g., in a read-modify-write), the parity is checked as part of the read operations. PCOR based DirectProtect can have an effect on performance because every read and write involves every data member of the RAID set this performance impact varies with data access patterns. An I/O pattern in which every I/O is full-stripealigned (the I/O size equals the stripe size and is aligned on a stripe boundary) naturally involves every data member of the RAID set and has minimal performance impact with PCOR DirectProtect on. Sequential I/O patterns, in which the read-ahead or write-back cache can turn non-stripe-aligned I/Os into full-stripe aligned RAID set I/Os, have minimal performance impact with PCOR DirectProtect on. Small random reads performed with PCOR DirectProtect enabled (reads that access fewer disks than the number in a full stripe) will suffer more degradation due to the requirement to read from all the disks in the stripe to check parity. 2 In computer science, ACID (Atomicity, Consistency, Isolation, Durability) is a set of properties that guarantee that database transactions are processed reliably. In the context of databases, a single logical operation on the data is called a transaction. An example of a transaction is a transfer of funds from one bank account to another, even though it might consist of multiple individual operations (such as debiting one account and crediting another). This brief definition was obtained from Wikipedia: 10

11 Data Integrity Field DirectProtect: This approach to detecting and correcting Physical Disk errors stores redundant information about the data in a form other than RAID parity. One approach is to store a hash (e.g., a CRC check) or Data Integrity Field (DIF) of each block s data, and then check this each time the block is read. Of course, the physical disk already stores a sophisticated Reed-Solomon code for each block that both detects and corrects errors, so having the RAID system store another hash may seem redundant but remember that the purpose of SFA DirectProtect is to improve the undetected errors of low-cost physical disks. There are several advantages to performing data integrity verification with a DIF vs. the PCOR method alone. The first is that calculating hash is far less intensive than calculating parity and hence results in significantly smaller levels of performance degradation. Additionally, the DIF method can easily detect and correct silent data corruption on mirrored (RAID 1) raid sets. Lastly, the DIF method has become accepted and standardized in the form of ANSI T10- DIF. This means it may be possible in a future version of SFA OS to emulate complete end-to-end data integrity checks even with SATA disk drives. To improve DirectProtect performance and provide additional data integrity checking, SFA OS includes this secondary DIF method for ensuring data integrity on SATA disks. This method inserts an additional 512-byte DIF block on each physical disk for every 64 data blocks on that disk which is used to store a hash of the data in each of those 64 blocks. When data is read from a physical disk, the read is lengthened to include the DIF block and the hash code is calculated and checked against the value stored in the DIF block. If an error is detected then steps will be taken to correct the error using retries and RAID redundancy. DIF blocks are cached to minimize the impact on performance. Storage System Efficiencies Storage Fusion Xcelerator (SFX) SFX is a suite of storage acceleration tools that combine spinning and solid state disk storage media with application aware technology to intelligently accelerate block and file-based data access. It is part of the SFA Operating System and extends the functionality of the storage system s cache by selectively frontending traditional rotating media with some amount of flash memory. This yields acceleration in the context of the application. SFX consists of a pool of SSD flash-based drives that actually become an extension of the DRAM cache. SFX cache can be allocated to a Logical Unit Number (LUN), which refers to a logical disk created from a group of real disks, or can be shared between multiple LUNs. It has the effect of front-ending the LUN with some very fast and large cache, without having to dedicate expensive SSD drives to a single LUN. There are currently four modes of SFX cache, which determine how data is served, and when data gets flushed-out of cache based on available headroom. SFX Read Cache This mode is designed for read intensive workloads. It caches frequently accessed data sets in the faster SFX tier to significantly speed up application performance. 11

12 SFX Write Cache This mode is designed for write intensive workloads. It allows large writes to burst at full speed to the SFX tier - and then groups the data and writes it down to disk over time eliminating the need to deploy petabytes of rotating media to meet extreme performance requirements. SFX Instant Commit This mode is designed for read after writes. It populates the SFX tier with new writes to warmup cache as well as accelerating subsequent reads. SFX Content Commit This mode allows applications and file systems to send down hints to the storage system, delivering the best storage acceleration possible by eliminating the need to guess what the IO pattern will be upon deployment. More details on SFX is available in the DDN white paper titled Storage Fusion Xcelerator. Partial Disk Rebuild When a disk does fail, the SFA OS tracks the changes made to a RAID set when a member physical disk becomes unavailable, and if that member becomes available again within a user-settable timeout then only the stripes that were modified while the member was missing are rebuilt. This minimizes the mean-time-to-repair for the RAID set and thus improves the data reliability of the RAID set while also limiting any performance impact of a drive repair. Real-time Adaptive Cache Technology (ReACT ) Because the SFA12KX performs at extreme levels in both IOPS and bandwidth, it was desirable to achieve extreme performance in mixed workload scenarios. Given a logical unit, where data I/O is comprised of both random I/O and sequential I/O, it is desirable to enable caching (and cache mirroring) for high IOPS performance. With cache mirroring enabled, sequential I/O performance suffers by having to cross the inter-controller link. It also has the side effect of invalidating random I/O cache as it fills the cache and displaces previously cached data. To remedy this situation, SFA OS employs the ReACT feature to intelligently cache or write-through data based on incoming write patterns. With write-back cache enabled and ReACT disabled, the data written to a given logical disk with aligned full-stripe writes is cached in the write-back cache and mirrored to the partner controller. With ReACT enabled for a given pool, the data written to the pool with aligned full-stripe writes is not cached and is instead written directly to the physical disks (i.e., writethrough). Either way, non-aligned writes are written to write-back cache (Figure 5). By enabling ReACT, applications that generate aligned full-stripe writes can achieve higher performance because write data is not cached and thus is not mirrored, resulting in greatly reduced inter-controller link traffic. 12

13 Aligned I/O Single-Operation Parallelized Striped-Writes No Cache Mirroring Required for Fast Data P Q Unaligned I/O Write-Back Cache Mirrored Accelerated Write Performance Avoids RMW Performance Avoids RMW Performanc M Cache Mirror Figure 5 Optimizing Cache Utilization with ReACT Rebuild Priority SFA OS employs a tunable parameter per pool for rebuild priority. Adjusting this setting will cause the rebuild engine to use less or more system resources for the rebuild operation. This feature allows an administrator the flexibility to adjust rebuild priority in relation to overall system performance. A lower rebuild priority setting will consume less of these system resources which will allow the system to devote more resources to incoming I/O. Conversely, it may be appropriate to increase rebuild priority to shorten rebuild time. Read Quality of Service DDN provides a highly reliable quality of service on read that allows the SFA12KX to stream data with very low latency. This allows latency sensitive application such as video streaming to deliver consistent and predictable performance even during system component failures. 13

14 Management DirectMon Today, organizations are facing an exponential increase in the amount of data being created. The ability to successfully manage this data, coupled with the growing complexity of storage infrastructures is creating significant challenges for IT managers. While the cost of maintaining storage infrastructures continues to increase, headcount and budget remains fixed. What is needed is an advanced management platform that reduces the cost and complexity of storage management. DirectMon is an advanced configuration and monitoring solution that leverages our leadership in supporting the world s largest file storage systems for over a decade. Purpose-built to improve the performance of IT operations, top-down support is provided for managing multiple DDN SFA storage arrays including the SFA12KX, SFA10K and the GRIDScaler and EXAScaler clustered file system appliances. Removing the complexity out of managing storage, its ease-of-use features and notifications allow administrators to quickly resolve problems, freeing-up valuable time to concentrate on more important tasks. DirectMon is ideally suited for any size IT environment to help simplify the configuration and management of the storage infrastructure, even as data continues to grow exponentially. Application Programming Interface (API) In addition to the traditional ways of configuring and managing the SFA storage products via CLI, GUI and SNMP interfaces, API s are provided to give customers the ability to programmatically configure and manage the SFA storage products. Enterprises will now be able to integrate our SFA products into their overall management framework. Python based API clients are provided to simplify the integration effort. 14

15 Software Summary SFA OS on the SFA12KX introduces new levels of extreme performance with several unique features. The scalable architecture is also very expandable. The SFA12KX with SFA OS provides the foundation upon which new data management and block virtualization features will be built in future releases. The flexibility and architecture of SFA OS allows these new features to be developed quickly, with rapid evolution of features utilizing the same hardware platform. This evolution will provide long-term investment protection and enhances longevity of SFA based products. SFA12KX Hardware Architecture The last several years have seen significant improvements in multiple commodity components. As mentioned previously, processors are increasing in the number of computing cores and the speed of those cores. The processor and bus interconnects have evolved to speeds that were only available with proprietary designs just a short time ago. HyperTransport (HT) and Intel QuickPath Interconnect (QPI) have replaced slow Front Side Bus (FSB) technology with low latency, point-topoint links featuring revolutionary bi-directional transfer speeds. Now that both AMD and Intel have adopted the practice of integrating the memory controllers, memory access speeds are greatly increased and experience lower latency. Peripheral buses have converged to PCI-Express (PCIe) which is now in its 3rd generation, and processors now have multiple integrated PCIe interfaces. Thus, nearly all the major components, busses and I/O paths around commodity computing processors have greatly improved in just the last couple of years. Combining these processing and I/O capabilities with current HBAs, HCAs and NICs in a unique configuration yields an extremely powerful storage hardware platform (Figure 2). RAID Processing A powerful storage hardware platform is useless without a tightly integrated software architecture that squeezes every bit of performance from the components and makes them work in a harmonious fashion. The SFA data integrity engine has been written from the ground up to be multi-threaded and highly parallelized to take maximum advantage of multi-core, multi-thread storage processors. Not only do various elements of the RAID stack run in parallel but there are two parallel instances of the storage engine: one in RAID processor (RP)0 and one in RP1 (Figure 6). Thus, the SFA12KX actually has two parallel, multi-threaded RAID engines that work simultaneously in each controller for a total of 4 RAID processors across the redundant controller pair. Further, each RAID processor runs multiple threads that manage the SFA cache, data integrity calculations and I/O movers. Thus, as the number of storage system cores are increased, additional parallel processes can be run simultaneously and both IOPS and bandwidth will increase accordingly. I/O Channels and Architecture 15

16 Powerful parallel RAID processors need to be able to handle massive amounts of I/O on the front end (block interfaces) and the back end (disk drive pool). The SFA12KX meets this challenge by providing each RAID processor with its own dedicated I/O channels to Fibre Channel or InfiniBand host interfaces on the front end to balance performance to SAS disk-enclosure interfaces on the back end. A very high speed, low latency interconnect allows for data transfers between RAID processors if and when necessary. This arrangement allows the SFA12KX to perform at extreme data rates as data is streamed from the host interfaces directly into RAID processors and out the back end to disks without having to contend for a shared I/O bus. The ability to move data through the controller in a streamlined fashion is what also gives the SFA12KX the ability to perform at extreme levels in IOPS. The ability to communicate via an unprecedented number of channels across multiple disks simultaneously is what allows the SFA12KX to achieve over 1.4 million sustained IOPS to SSD drives. Front End Block Interfaces, Up to 160 Gb/s Controller 0 SFA12KX IO Paths 240 Gb/s Front End Block Interfaces, Up to 160 Gb/s Controller 1 RPO High Speed, RP1 Inter-Controller Links RPO RP1 20 6Gb/s SAS x4 Links, 480Gb/s Up to 1680 SAS, SATA or SSD Drives Figure 6 SFA12KX Streamlined I/O Paths 20 6Gb/s SAS x4 Links, 480Gb/s Cache Extreme IOPS performance to disk is important but for small size, high IOPS data patterns where latency becomes the gating factor, cache is a necessity. The SFA12KX offers high levels of mirrored cache, 32 GB total. Cache is implemented in DDR3 SDRAM memory for the lowest latency, highest performing cache. In the case of a power event, SFA12KX utilizes a dedicated battery backup unit to hold up the controller while the un-flushed write-back cache data is transferred to internal, nonvolatile, mirrored storage. Back End Disk Connectivity Overall, the design of the SFA12KX hardware is about balance. The extreme performance capabilities of the host ports are facilitated by a streamlined I/O path directly to the back end disks. The massive 960Gb/s internal SAS network not only serves the IOPS and bandwidth needs of the controller itself, but has ample headroom for additional I/O operations internal to the architecture. This headroom allows disk rebuild I/O to coexist with application service as there is plenty of bandwidth for both to occur simultaneously. By providing 40 x 4 SAS channels to serve 1,680 disk drives, the ratio of drives per channel is decreased. This arrangement allows for more commands to be queued per drive as well as providing ample bandwidth for 16

17 high IOPS SSD drives. Additionally, because all of the disk enclosures are best-practice configured either as 5, 10 or 20 enclosures per SFA couplet the SFA12KX has the ability to RAID storage enclosures for high levels of enclosure fault tolerance. Using an 8+2 RAID 6 configuration, the SFA controller can lose up to 4 drive enclosures (or 2/10ths of the system resources) on an active system and still deliver full access to online data. Hardware Summary This unique combination of high-performance storage processing technologies, married to an advanced, optimized software architecture not only make the SFA12KX the leader in IOPS and bandwidth, but more importantly, serves as a high-density, fault-tolerant storage foundation for evolutionary SFA OS advances far into the future. SFA OS forms the basis for the next generation of ultra-high performance block storage. This unique hardware and software combination also lends itself to more interesting possibilities, further differentiating SFA OS. SFA OS and In-Storage Processing Computing Systems The decision to marry unique and specialized software to industry standard hardware components in SFA lends itself to an innovation that goes far beyond block storage services. SFA OS allows for embedding applications within the SFA12KXE. The applications that make the most sense to embed (initially) are those that would benefit the most from reduced latency and high bandwidth: clustered file system services. Thus, in its first iteration the SFA12KXE has the capability to embed the Lustre file system (the OSSs) or IBM GPFS (the NSDs). Embedding the file system servers within the storage device reduces the number of servers, infrastructure requirements and network connections which in turn reduces complexity, power consumption and cooling requirements. At the same time, it streamlines I/O and reduces latency by removing data hops and eliminates wasteful storage protocol conversion. In-Storage Processing Computing Capability SFA OS utilizes virtualization software to allow for applications to be run inside the storage device. Various methods of memory and resource protection are employed to guard the block RAID functionality and ensure overall system resources are allocated in a secure and controlled fashion. SFA OS acts as a hypervisor, using technologies such as ccnuma and KVM to control processor, core, memory, I/O and virtual disk allocations. This ensures that applications that run in the embedded space cannot affect the block RAID process memory space and that the applications only utilize processing and I/O resources they have been assigned. Virtualization technologies are usually associated with performance degradation, not improvements in performance. Though SFA OS uses software and hardware virtualization, special care and development have been undertaken to ensure not only as little performance degradation as possible, but to produce an environment that offers enhanced performance. This is largely achieved with two distinct methods. PCIe Device Dedication 17

18 In the case of Lustre and GPFS, Infiniband or Ethernet HCAs are commonly used as the frontend interfaces to the file system servers. Normally, virtualization technologies share hardware devices such as HCA s among virtual machines; slowing access for all and requiring virtual device drivers. SFA overcomes these traditional bottlenecks by dedicating PCIe devices directly to virtual machines. In the course of virtual machine initialization, the PCIe address space for the PCIe device in question is remapped to the virtual machine space. When the virtual machine boots its associated OS, it sees the PCIe device (in this case, the Infiniband or Ethernet card) natively, as if it was running on a physical machine. This allows the use of the HCA s native software drivers, eliminating any need for a virtual device. Utilizing this method, virtual machines running inside the SFA12KXE have been able to achieve external bandwidth of 20GB/s or more. Virtual Disk Driver By dedicating PCIe devices directly to virtual machines there is no need to modify OS images or supply highly specialized virtual I/O devices. Virtual machines running inside an SFA12KXE enjoy nearly native speed access to HCAs. The remaining hurdle is access to the virtual disks (LUNs) served by the block RAID services side of SFA from the OS running inside the virtual machine. This access is achieved with the addition of a small, lightweight kernel module to the Linux image running inside the virtual machine. This driver presents virtual disks assigned to the virtual machine as standard Linux block devices under /dev. What looks like a standard block device is actually a shared memory interface between the virtual machine and the block RAID services managed by SFA OS. As shown in Fig. 7, what was a dedicated server, an FC HBA, an FC switch and another FC HBA is reduced down to a direct memory interface at processor bus speeds. For writes from the OS to the device, data in memory is copied from the virtual machine space to the RAID space before it is manipulated by the RAID engine. This prevents the virtual machine from having write access to the RAID memory space. IO Paths Latency Traditional Client HCA/NIC Switch HCA/NIC Server HBA SAN Switch HBA Storage Components Eliminated withsfa12k In-Storage Computing SFA Embedded Applications Client HCA/NIC Switch HCA/NIC Application Efficiently Place Data Directly Into SFA Memory Elimination of Protocol Conversion Reduces Latency, Improves IOPS Performance Figure 7 I/O Path Reduct/n in SFA12KX In-Storage Computing Systems On reads of the virtual disk device, the block RAID engine reads from disk, places the data in memory and passes a shared pointer to the virtual disk driver so that the virtual machine can read directly from the RAID engine without a memory copy. 18

19 Thus, IOPS intensive loads (such as file system metadata operations) can enjoy greatly reduced latency. The removal of SCSI protocol overhead, Fibre Channel interconnects, SAN switches and interface conversion reduces storage response times and lets the embedded file system take full advantage of the SFA12KXE high performance random I/O capabilities. This I/O streamlining in turn improves performance for transaction intensive workloads. Reduction in Equipment, Infrastructure and Complexity By combining virtualization, an advanced new block RAID architecture and cutting edge hardware technology it s possible to achieve high performance while at the same time reducing complexity. As shown in Fig. 8, using the Lustre file system as an example, SFA technology can result in as much as a 10 to 1 reduction in the number of managed systems depending on deployment. While clustered file system services were the first choice for applications to be embedded within SFA, virtually any application that would benefit from ultra-low latency access to block disk devices could benefit from being embedded. As processors increase in speed and the number of cores, the possibilities for what can be embedded increase along with the performance of the block RAID engine. Traditional Lustre Deployment to Achieve 5 GB/s SFA12KXE, Embedded EXA Scaler Lustre Clients IB or 10 Gig-E Lustre MGS Lustre Clients IB or 10 Gig-E Lustre OSS Nodes SAN Fibre Channel Lustre MDS Nodes Active Fibre Channel Standby External RAID Array Disk Drives Disk Drives 10 managed systems: 2+ RAID Arrays 7 Servers 1 Fibre Switch 1 SFA12KXE Storage Building Block Figure 8 Reduction in Equipment, Infrastructure and Complexity with SFA12KXE 19

20 SFA12KX Family: Summary Disk storage systems simply enable computational output to reside on non-volatile media as opposed to being dependent on more volatile media (RAM). Thus, its purpose is to serve compute clients rapidly with performance and integrity predictability. To the storage environment, it should not matter if systems are processing data for a Fortune 500 enterprise, climatologists predicting global weather patterns or scientists simulating high-energy physics. What does matter is that the technology used in those computers is becoming ever more multi-threaded. The resulting effect on storage systems is the simultaneous read and write of multiple files whose access histogram is seen as mixed or highly transactional to the supporting storage systems. Thus, storage systems must adapt to changing data patterns to accommodate serving multithreaded compute clients without bottlenecking application I/O. SFA12KX meets the challenges of changing data patterns by offering extreme performance in both IOPS and bandwidth. A unique combination of an entirely new storage operating system (SFA OS) and best of breed storage processing components have made a system architecture that performs well at both ends of the I/O spectrum a reality. In addition to meeting the mixed I/O requirements of the most intensive compute environments, SFA OS also allows for embedding clustered file system services directly inside the block storage device. This capability results in the reduction of servers, infrastructure and complexity. In addition to reducing the complexity of scale-out storage, Storage Fusion Architecture can also increase storage responsiveness by removing latency-injecting elements from the storage cluster. Now that DDN s move to high speed storage processing systems is complete, rapid development of additional features is possible: advanced storage virtualization capabilities, data management features and advanced application encapsulation, resulting in infrastructure and complexity reduction. The SFA12KX family is the leader in performance (in both IOPS and bandwidth) but Storage Fusion Architecture ensures enduring leadership as it readily adapts to and benefits from advances in the processing components it utilizes. DDN About Us DataDirect Networks (DDN) is the world s largest privately held information storage company. We are the leading provider of data storage and processing solutions and services, that enable content-rich and high growth IT environments to achieve the highest levels of systems scalability, efficiency and simplicity. DDN enables enterprises to extract value and deliver results from their information. Our customers include the world s leading online content and social networking providers, high performance cloud and grid computing, life sciences, media production organizations and security & intelligence organizations. Deployed in thousands of mission critical environments worldwide, DDN s solutions have been designed, engineered and proven in the world s most scalable data centers, to ensure competitive business advantage for today s information powered enterprise. For more information, go to www. or call DataDirect Networks, Inc. All Rights Reserved. Storage Fusion Architecture, Storage Fusion Xcelerator, DirectRAID, DirectProtect, EXAScaler, GRIDScaler, In-Storage Processing, ReACT, SFA10K, SFA12KE, S2A, SFX are trademarks of DataDirect Networks. All other trademarks are the property of their respective owners. 20

SFA10K-X & SFA10K-E. ddn.com. Storage Fusion Architecture TM. DDN Whitepaper

SFA10K-X & SFA10K-E. ddn.com. Storage Fusion Architecture TM. DDN Whitepaper DDN Whitepaper Storage Fusion Architecture TM SFA10K-X & SFA10K-E Breaking Down Storage Barriers by Providing Extreme Performance in Both Bandwidth and IOPS Table of Contents Abstract 3 Introduction 3

More information

ANY SURVEILLANCE, ANYWHERE, ANYTIME

ANY SURVEILLANCE, ANYWHERE, ANYTIME ANY SURVEILLANCE, ANYWHERE, ANYTIME WHITEPAPER DDN Storage Powers Next Generation Video Surveillance Infrastructure INTRODUCTION Over the past decade, the world has seen tremendous growth in the use of

More information

HadoopTM Analytics DDN

HadoopTM Analytics DDN DDN Solution Brief Accelerate> HadoopTM Analytics with the SFA Big Data Platform Organizations that need to extract value from all data can leverage the award winning SFA platform to really accelerate

More information

With DDN Big Data Storage

With DDN Big Data Storage DDN Solution Brief Accelerate > ISR With DDN Big Data Storage The Way to Capture and Analyze the Growing Amount of Data Created by New Technologies 2012 DataDirect Networks. All Rights Reserved. The Big

More information

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010 Flash Memory Arrays Enabling the Virtualized Data Center July 2010 2 Flash Memory Arrays Enabling the Virtualized Data Center This White Paper describes a new product category, the flash Memory Array,

More information

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com

Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...

More information

The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage

The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage sponsored by Dan Sullivan Chapter 1: Advantages of Hybrid Storage... 1 Overview of Flash Deployment in Hybrid Storage Systems...

More information

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies

More information

EMC XTREMIO EXECUTIVE OVERVIEW

EMC XTREMIO EXECUTIVE OVERVIEW EMC XTREMIO EXECUTIVE OVERVIEW COMPANY BACKGROUND XtremIO develops enterprise data storage systems based completely on random access media such as flash solid-state drives (SSDs). By leveraging the underlying

More information

Storage Fusion Xcelerator

Storage Fusion Xcelerator Storage Fusion Xcelerator DDN Whitepaper Abstract: As organizations look for ways to increase the performance of their Big Data applications and file systems, they are increasingly turning to solid state

More information

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

WHITEPAPER: Understanding Pillar Axiom Data Protection Options WHITEPAPER: Understanding Pillar Axiom Data Protection Options Introduction This document gives an overview of the Pillar Data System Axiom RAID protection schemas. It does not delve into corner cases

More information

Using Multipathing Technology to Achieve a High Availability Solution

Using Multipathing Technology to Achieve a High Availability Solution Using Multipathing Technology to Achieve a High Availability Solution Table of Contents Introduction...3 Multipathing Technology...3 Multipathing I/O Implementations...5 Storage Redundancy...5 Infortrend

More information

Architecting a High Performance Storage System

Architecting a High Performance Storage System WHITE PAPER Intel Enterprise Edition for Lustre* Software High Performance Data Division Architecting a High Performance Storage System January 2014 Contents Introduction... 1 A Systematic Approach to

More information

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Built up on Cisco s big data common platform architecture (CPA), a

More information

SMB Direct for SQL Server and Private Cloud

SMB Direct for SQL Server and Private Cloud SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server

More information

New Storage System Solutions

New Storage System Solutions New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems

More information

Object storage in Cloud Computing and Embedded Processing

Object storage in Cloud Computing and Embedded Processing Object storage in Cloud Computing and Embedded Processing Jan Jitze Krol Systems Engineer DDN We Accelerate Information Insight DDN is a Leader in Massively Scalable Platforms and Solutions for Big Data

More information

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute PADS GPFS Filesystem: Crash Root Cause Analysis Computation Institute Argonne National Laboratory Table of Contents Purpose 1 Terminology 2 Infrastructure 4 Timeline of Events 5 Background 5 Corruption

More information

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014 Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,

More information

HP Smart Array Controllers and basic RAID performance factors

HP Smart Array Controllers and basic RAID performance factors Technical white paper HP Smart Array Controllers and basic RAID performance factors Technology brief Table of contents Abstract 2 Benefits of drive arrays 2 Factors that affect performance 2 HP Smart Array

More information

DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering

DELL RAID PRIMER DELL PERC RAID CONTROLLERS. Joe H. Trickey III. Dell Storage RAID Product Marketing. John Seward. Dell Storage RAID Engineering DELL RAID PRIMER DELL PERC RAID CONTROLLERS Joe H. Trickey III Dell Storage RAID Product Marketing John Seward Dell Storage RAID Engineering http://www.dell.com/content/topics/topic.aspx/global/products/pvaul/top

More information

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks

Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance

More information

Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage. White Paper

Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage. White Paper Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage White Paper June 2011 2011 Coraid, Inc. Coraid, Inc. The trademarks, logos, and service marks (collectively "Trademarks") appearing on the

More information

UCS M-Series Modular Servers

UCS M-Series Modular Servers UCS M-Series Modular Servers The Next Wave of UCS Innovation Marian Klas Cisco Systems June 2015 Cisco UCS - Powering Applications at Every Scale Edge-Scale Computing Cloud-Scale Computing Seamlessly Extend

More information

A Survey of Shared File Systems

A Survey of Shared File Systems Technical Paper A Survey of Shared File Systems Determining the Best Choice for your Distributed Applications A Survey of Shared File Systems A Survey of Shared File Systems Table of Contents Introduction...

More information

SAN Conceptual and Design Basics

SAN Conceptual and Design Basics TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer

More information

PARALLELS CLOUD STORAGE

PARALLELS CLOUD STORAGE PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...

More information

Unisys ClearPath Forward Fabric Based Platform to Power the Weather Enterprise

Unisys ClearPath Forward Fabric Based Platform to Power the Weather Enterprise Unisys ClearPath Forward Fabric Based Platform to Power the Weather Enterprise Introducing Unisys All in One software based weather platform designed to reduce server space, streamline operations, consolidate

More information

Pivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview

Pivot3 Desktop Virtualization Appliances. vstac VDI Technology Overview Pivot3 Desktop Virtualization Appliances vstac VDI Technology Overview February 2012 Pivot3 Desktop Virtualization Technology Overview Table of Contents Executive Summary... 3 The Pivot3 VDI Appliance...

More information

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer

The Future of Computing Cisco Unified Computing System. Markus Kunstmann Channels Systems Engineer The Future of Computing Cisco Unified Computing System Markus Kunstmann Channels Systems Engineer 2009 Cisco Systems, Inc. All rights reserved. Data Centers Are under Increasing Pressure Collaboration

More information

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org

IOmark- VDI. Nimbus Data Gemini Test Report: VDI- 130906- a Test Report Date: 6, September 2013. www.iomark.org IOmark- VDI Nimbus Data Gemini Test Report: VDI- 130906- a Test Copyright 2010-2013 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VDI, VDI- IOmark, and IOmark are trademarks of Evaluator

More information

Big data management with IBM General Parallel File System

Big data management with IBM General Parallel File System Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers

More information

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system Christian Clémençon (EPFL-DIT)  4 April 2013 GPFS Storage Server Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " Agenda" GPFS Overview" Classical versus GSS I/O Solution" GPFS Storage Server (GSS)" GPFS Native RAID

More information

High Availability with Windows Server 2012 Release Candidate

High Availability with Windows Server 2012 Release Candidate High Availability with Windows Server 2012 Release Candidate Windows Server 2012 Release Candidate (RC) delivers innovative new capabilities that enable you to build dynamic storage and availability solutions

More information

Virtualization of the MS Exchange Server Environment

Virtualization of the MS Exchange Server Environment MS Exchange Server Acceleration Maximizing Users in a Virtualized Environment with Flash-Powered Consolidation Allon Cohen, PhD OCZ Technology Group Introduction Microsoft (MS) Exchange Server is one of

More information

ECMWF HPC Workshop: Accelerating Data Management

ECMWF HPC Workshop: Accelerating Data Management October 2012 ECMWF HPC Workshop: Accelerating Data Management Massively-Scalable Platforms and Solutions Engineered for the Big Data and Cloud Era Glenn Wright Systems Architect, DDN Data-Driven Paradigm

More information

Oracle Database Scalability in VMware ESX VMware ESX 3.5

Oracle Database Scalability in VMware ESX VMware ESX 3.5 Performance Study Oracle Database Scalability in VMware ESX VMware ESX 3.5 Database applications running on individual physical servers represent a large consolidation opportunity. However enterprises

More information

IBM Global Technology Services September 2007. NAS systems scale out to meet growing storage demand.

IBM Global Technology Services September 2007. NAS systems scale out to meet growing storage demand. IBM Global Technology Services September 2007 NAS systems scale out to meet Page 2 Contents 2 Introduction 2 Understanding the traditional NAS role 3 Gaining NAS benefits 4 NAS shortcomings in enterprise

More information

Fault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems

Fault Tolerance & Reliability CDA 5140. Chapter 3 RAID & Sample Commercial FT Systems Fault Tolerance & Reliability CDA 5140 Chapter 3 RAID & Sample Commercial FT Systems - basic concept in these, as with codes, is redundancy to allow system to continue operation even if some components

More information

Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation

Overview of I/O Performance and RAID in an RDBMS Environment. By: Edward Whalen Performance Tuning Corporation Overview of I/O Performance and RAID in an RDBMS Environment By: Edward Whalen Performance Tuning Corporation Abstract This paper covers the fundamentals of I/O topics and an overview of RAID levels commonly

More information

Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. ddn.com

Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. ddn.com DDN Technical Brief Modernizing Hadoop Architecture for Superior Scalability, Efficiency & Productive Throughput. A Fundamentally Different Approach To Enterprise Analytics Architecture: A Scalable Unit

More information

TECHNOLOGY BRIEF. Compaq RAID on a Chip Technology EXECUTIVE SUMMARY CONTENTS

TECHNOLOGY BRIEF. Compaq RAID on a Chip Technology EXECUTIVE SUMMARY CONTENTS TECHNOLOGY BRIEF August 1999 Compaq Computer Corporation Prepared by ISSD Technology Communications CONTENTS Executive Summary 1 Introduction 3 Subsystem Technology 3 Processor 3 SCSI Chip4 PCI Bridge

More information

Fibre Forward - Why Storage Infrastructures Should Be Built With Fibre Channel

Fibre Forward - Why Storage Infrastructures Should Be Built With Fibre Channel Fibre Forward - Why Storage Infrastructures Should Be Built With Fibre Channel Prepared by: George Crump, Lead Analyst Prepared: June 2014 Fibre Forward - Why Storage Infrastructures Should Be Built With

More information

The Benefit of Migrating from 4Gb to 8Gb Fibre Channel

The Benefit of Migrating from 4Gb to 8Gb Fibre Channel The Benefit of Migrating from 4Gb to 8Gb Fibre Channel Notices The information in this document is subject to change without notice. While every effort has been made to ensure that all information in this

More information

Taking Linux File and Storage Systems into the Future. Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated

Taking Linux File and Storage Systems into the Future. Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated Taking Linux File and Storage Systems into the Future Ric Wheeler Director Kernel File and Storage Team Red Hat, Incorporated 1 Overview Going Bigger Going Faster Support for New Hardware Current Areas

More information

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance. Agenda Enterprise Performance Factors Overall Enterprise Performance Factors Best Practice for generic Enterprise Best Practice for 3-tiers Enterprise Hardware Load Balancer Basic Unix Tuning Performance

More information

FLASH 15 MINUTE GUIDE DELIVER MORE VALUE AT LOWER COST WITH XTREMIO ALL- FLASH ARRAY Unparal eled performance with in- line data services al the time

FLASH 15 MINUTE GUIDE DELIVER MORE VALUE AT LOWER COST WITH XTREMIO ALL- FLASH ARRAY Unparal eled performance with in- line data services al the time FLASH 15 MINUTE GUIDE DELIVER MORE VALUE AT LOWER COST WITH XTREMIO ALL- FLASH ARRAY Unparalleled performance with in- line data services all the time OVERVIEW Opportunities to truly innovate are rare.

More information

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one

More information

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Using Synology SSD Technology to Enhance System Performance Synology Inc. Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...

More information

Technology Insight Series

Technology Insight Series Evaluating Storage Technologies for Virtual Server Environments Russ Fellows June, 2010 Technology Insight Series Evaluator Group Copyright 2010 Evaluator Group, Inc. All rights reserved Executive Summary

More information

Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers

Improving Time to Results for Seismic Processing with Paradigm and DDN. ddn.com. DDN Whitepaper. James Coomer and Laurent Thiers DDN Whitepaper Improving Time to Results for Seismic Processing with Paradigm and DDN James Coomer and Laurent Thiers 2014 DataDirect Networks. All Rights Reserved. Executive Summary Companies in the oil

More information

Q & A From Hitachi Data Systems WebTech Presentation:

Q & A From Hitachi Data Systems WebTech Presentation: Q & A From Hitachi Data Systems WebTech Presentation: RAID Concepts 1. Is the chunk size the same for all Hitachi Data Systems storage systems, i.e., Adaptable Modular Systems, Network Storage Controller,

More information

IBM System x GPFS Storage Server

IBM System x GPFS Storage Server IBM System x GPFS Storage Crispin Keable Technical Computing Architect 1 IBM Technical Computing comprehensive portfolio uniquely addresses supercomputing and mainstream client needs Technical Computing

More information

MS Exchange Server Acceleration

MS Exchange Server Acceleration White Paper MS Exchange Server Acceleration Using virtualization to dramatically maximize user experience for Microsoft Exchange Server Allon Cohen, PhD Scott Harlin OCZ Storage Solutions, Inc. A Toshiba

More information

General Parallel File System (GPFS) Native RAID For 100,000-Disk Petascale Systems

General Parallel File System (GPFS) Native RAID For 100,000-Disk Petascale Systems General Parallel File System (GPFS) Native RAID For 100,000-Disk Petascale Systems Veera Deenadhayalan IBM Almaden Research Center 2011 IBM Corporation Hard Disk Rates Are Lagging There have been recent

More information

The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays

The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays Executive Summary Microsoft SQL has evolved beyond serving simple workgroups to a platform delivering sophisticated

More information

SOLID STATE DRIVES AND PARALLEL STORAGE

SOLID STATE DRIVES AND PARALLEL STORAGE SOLID STATE DRIVES AND PARALLEL STORAGE White paper JANUARY 2013 1.888.PANASAS www.panasas.com Overview Solid State Drives (SSDs) have been touted for some time as a disruptive technology in the storage

More information

IBM ^ xseries ServeRAID Technology

IBM ^ xseries ServeRAID Technology IBM ^ xseries ServeRAID Technology Reliability through RAID technology Executive Summary: t long ago, business-critical computing on industry-standard platforms was unheard of. Proprietary systems were

More information

ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter

ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter WHITEPAPER ANY THREAT, ANYWHERE, ANYTIME Scalable.Infrastructure.to.Enable.the.Warfi.ghter THE BIG DATA CHALLENGE AND OPPORTUNITY The.proliferation,.management.and.analysis.of.intelligence.data.is.a.fast.growing.concern.

More information

Benchmarking Cassandra on Violin

Benchmarking Cassandra on Violin Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract

More information

ioscale: The Holy Grail for Hyperscale

ioscale: The Holy Grail for Hyperscale ioscale: The Holy Grail for Hyperscale The New World of Hyperscale Hyperscale describes new cloud computing deployments where hundreds or thousands of distributed servers support millions of remote, often

More information

How To Connect Virtual Fibre Channel To A Virtual Box On A Hyperv Virtual Machine

How To Connect Virtual Fibre Channel To A Virtual Box On A Hyperv Virtual Machine Virtual Fibre Channel for Hyper-V Virtual Fibre Channel for Hyper-V, a new technology available in Microsoft Windows Server 2012, allows direct access to Fibre Channel (FC) shared storage by multiple guest

More information

RAID technology and IBM TotalStorage NAS products

RAID technology and IBM TotalStorage NAS products IBM TotalStorage Network Attached Storage October 2001 RAID technology and IBM TotalStorage NAS products By Janet Anglin and Chris Durham Storage Networking Architecture, SSG Page No.1 Contents 2 RAID

More information

Understanding Microsoft Storage Spaces

Understanding Microsoft Storage Spaces S T O R A G E Understanding Microsoft Storage Spaces A critical look at its key features and value proposition for storage administrators A Microsoft s Storage Spaces solution offers storage administrators

More information

Intel RAID Controllers

Intel RAID Controllers Intel RAID Controllers Best Practices White Paper April, 2008 Enterprise Platforms and Services Division - Marketing Revision History Date Revision Number April, 2008 1.0 Initial release. Modifications

More information

The Data Placement Challenge

The Data Placement Challenge The Data Placement Challenge Entire Dataset Applications Active Data Lowest $/IOP Highest throughput Lowest latency 10-20% Right Place Right Cost Right Time 100% 2 2 What s Driving the AST Discussion?

More information

EMC XtremSF: Delivering Next Generation Performance for Oracle Database

EMC XtremSF: Delivering Next Generation Performance for Oracle Database White Paper EMC XtremSF: Delivering Next Generation Performance for Oracle Database Abstract This white paper addresses the challenges currently facing business executives to store and process the growing

More information

June 2009. Blade.org 2009 ALL RIGHTS RESERVED

June 2009. Blade.org 2009 ALL RIGHTS RESERVED Contributions for this vendor neutral technology paper have been provided by Blade.org members including NetApp, BLADE Network Technologies, and Double-Take Software. June 2009 Blade.org 2009 ALL RIGHTS

More information

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads 89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com @EdisonGroupInc 212.367.7400 IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads A Competitive Test and Evaluation Report

More information

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance Alex Ho, Product Manager Innodisk Corporation Outline Innodisk Introduction Industry Trend & Challenge

More information

Quantum StorNext. Product Brief: Distributed LAN Client

Quantum StorNext. Product Brief: Distributed LAN Client Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without

More information

Accelerating Applications and File Systems with Solid State Storage. Jacob Farmer, Cambridge Computer

Accelerating Applications and File Systems with Solid State Storage. Jacob Farmer, Cambridge Computer Accelerating Applications and File Systems with Solid State Storage Jacob Farmer, Cambridge Computer SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise

More information

SAS Business Analytics. Base SAS for SAS 9.2

SAS Business Analytics. Base SAS for SAS 9.2 Performance & Scalability of SAS Business Analytics on an NEC Express5800/A1080a (Intel Xeon 7500 series-based Platform) using Red Hat Enterprise Linux 5 SAS Business Analytics Base SAS for SAS 9.2 Red

More information

Violin Memory Arrays With IBM System Storage SAN Volume Control

Violin Memory Arrays With IBM System Storage SAN Volume Control Technical White Paper Report Best Practices Guide: Violin Memory Arrays With IBM System Storage SAN Volume Control Implementation Best Practices and Performance Considerations Version 1.0 Abstract This

More information

SUN STORAGE F5100 FLASH ARRAY

SUN STORAGE F5100 FLASH ARRAY SUN STORAGE F5100 FLASH ARRAY KEY FEATURES ACCELERATING DATABASE PERFORMANCE WITH THE WORLD S FASTEST SOLID- STATE FLASH ARRAY Unprecedented performance, power, and space efficiency World s first flash

More information

SanDisk ION Accelerator High Availability

SanDisk ION Accelerator High Availability WHITE PAPER SanDisk ION Accelerator High Availability 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Introduction 3 Basics of SanDisk ION Accelerator High Availability 3 ALUA Multipathing

More information

An Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine

An Oracle White Paper May 2011. Exadata Smart Flash Cache and the Oracle Exadata Database Machine An Oracle White Paper May 2011 Exadata Smart Flash Cache and the Oracle Exadata Database Machine Exadata Smart Flash Cache... 2 Oracle Database 11g: The First Flash Optimized Database... 2 Exadata Smart

More information

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications

More information

Cray DVS: Data Virtualization Service

Cray DVS: Data Virtualization Service Cray : Data Virtualization Service Stephen Sugiyama and David Wallace, Cray Inc. ABSTRACT: Cray, the Cray Data Virtualization Service, is a new capability being added to the XT software environment with

More information

Express5800 Scalable Enterprise Server Reference Architecture. For NEC PCIe SSD Appliance for Microsoft SQL Server

Express5800 Scalable Enterprise Server Reference Architecture. For NEC PCIe SSD Appliance for Microsoft SQL Server Express5800 Scalable Enterprise Server Reference Architecture For NEC PCIe SSD Appliance for Microsoft SQL Server An appliance that significantly improves performance of enterprise systems and large-scale

More information

Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era

Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era Enterprise Strategy Group Getting to the bigger truth. White Paper Direct Scale-out Flash Storage: Data Path Evolution for the Flash Storage Era Apeiron introduces NVMe-based storage innovation designed

More information

Application Performance for High Performance Computing Environments

Application Performance for High Performance Computing Environments Application Performance for High Performance Computing Environments Leveraging the strengths of Computationally intensive applications With high performance scale out file serving In data storage modules

More information

Introduction to Gluster. Versions 3.0.x

Introduction to Gluster. Versions 3.0.x Introduction to Gluster Versions 3.0.x Table of Contents Table of Contents... 2 Overview... 3 Gluster File System... 3 Gluster Storage Platform... 3 No metadata with the Elastic Hash Algorithm... 4 A Gluster

More information

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014

VMware Virtual SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Data Protection Advanced SEPTEMBER 2014 VMware SAN Backup Using VMware vsphere Table of Contents Introduction.... 3 vsphere Architectural Overview... 4 SAN Backup

More information

EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server

EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server White Paper EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server Abstract This white paper addresses the challenges currently facing business executives to store and process the growing

More information

21 st Century Storage What s New and What s Changing

21 st Century Storage What s New and What s Changing 21 st Century Storage What s New and What s Changing Randy Kerns Senior Strategist Evaluator Group Overview New technologies in storage - Continued evolution - Each has great economic value - Differing

More information

Server-Side Virtual Controller Technology (SVCT)

Server-Side Virtual Controller Technology (SVCT) TECHNOLOGY BRIEF Server-Side Virtual Controller Technology (SVCT) All traditional storage suffers from the I/O blender that severely impacts application performance, because traditional storage is still

More information

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre University of Cambridge, UIS, HPC Service Authors: Wojciech Turek, Paul Calleja, John Taylor

More information

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology

I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology I/O Virtualization Using Mellanox InfiniBand And Channel I/O Virtualization (CIOV) Technology Reduce I/O cost and power by 40 50% Reduce I/O real estate needs in blade servers through consolidation Maintain

More information

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5

More information

Windows 8 SMB 2.2 File Sharing Performance

Windows 8 SMB 2.2 File Sharing Performance Windows 8 SMB 2.2 File Sharing Performance Abstract This paper provides a preliminary analysis of the performance capabilities of the Server Message Block (SMB) 2.2 file sharing protocol with 10 gigabit

More information

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Using Synology SSD Technology to Enhance System Performance Synology Inc. Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_WP_ 20121112 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD

More information

Intel RAID SSD Cache Controller RCS25ZB040

Intel RAID SSD Cache Controller RCS25ZB040 SOLUTION Brief Intel RAID SSD Cache Controller RCS25ZB040 When Faster Matters Cost-Effective Intelligent RAID with Embedded High Performance Flash Intel RAID SSD Cache Controller RCS25ZB040 When Faster

More information

DDN in Seismic Workflows

DDN in Seismic Workflows DDN in Seismic Workflows October, 2014 Laura Shepard Director Vertical Markets Agenda About DDN Technology & Product Portfolio Technology Preview DDN About Us DDN is a Leader in Massively Scalable Platforms

More information

StorPool Distributed Storage Software Technical Overview

StorPool Distributed Storage Software Technical Overview StorPool Distributed Storage Software Technical Overview StorPool 2015 Page 1 of 8 StorPool Overview StorPool is distributed storage software. It pools the attached storage (hard disks or SSDs) of standard

More information

Building a Flash Fabric

Building a Flash Fabric Introduction Storage Area Networks dominate today s enterprise data centers. These specialized networks use fibre channel switches and Host Bus Adapters (HBAs) to connect to storage arrays. With software,

More information

All-Flash Arrays Weren t Built for Dynamic Environments. Here s Why... This whitepaper is based on content originally posted at www.frankdenneman.

All-Flash Arrays Weren t Built for Dynamic Environments. Here s Why... This whitepaper is based on content originally posted at www.frankdenneman. WHITE PAPER All-Flash Arrays Weren t Built for Dynamic Environments. Here s Why... This whitepaper is based on content originally posted at www.frankdenneman.nl 1 Monolithic shared storage architectures

More information

Hardware RAID vs. Software RAID: Which Implementation is Best for my Application?

Hardware RAID vs. Software RAID: Which Implementation is Best for my Application? STORAGE SOLUTIONS WHITE PAPER Hardware vs. Software : Which Implementation is Best for my Application? Contents Introduction...1 What is?...1 Software...1 Software Implementations...1 Hardware...2 Hardware

More information

STORAGE CENTER. The Industry s Only SAN with Automated Tiered Storage STORAGE CENTER

STORAGE CENTER. The Industry s Only SAN with Automated Tiered Storage STORAGE CENTER STORAGE CENTER DATASHEET STORAGE CENTER Go Beyond the Boundaries of Traditional Storage Systems Today s storage vendors promise to reduce the amount of time and money companies spend on storage but instead

More information

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products MaxDeploy Ready Hyper- Converged Virtualization Solution With SanDisk Fusion iomemory products MaxDeploy Ready products are configured and tested for support with Maxta software- defined storage and with

More information