Write Amplification: An Analysis of In-Memory Database Durability Techniques
|
|
|
- Jerome McCarthy
- 9 years ago
- Views:
Transcription
1 Write Amplification: An Analysis of In-Memory Database Durability Techniques Jaemyung Kim Cheriton School of Computer Science University of Waterloo Kenneth Salem Cheriton School of Computer Science University of Waterloo Khuzaima Daudjee Cheriton School of Computer Science University of Waterloo ABSTRACT Modern in-memory database systems perform transactions an order of magnitude faster than conventional database systems. While in-memory database systems can read the database without I/O, database updates can generate a substantial amount of I/O, since updates must normally be written to persistent secondary storage to ensure that they are durable. In this paper we present a study of storage managers for in-memory database systems, with the goal of characterizing their I/O efficiency. We model the storage efficiency of two classes of storage managers: those that perform in-place updates in secondary storage, and those that use copy-on-write. Our models allow us to make meaningful, quantitative comparisons of storage managers I/O efficiencies under a variety of conditions. 1. INTRODUCTION As memory prices continue to drop and memory sizes continue to increase, it has become possible to fit terabyte OLTP databases entirely in memory. Modern database systems can leverage these memory trends to deliver high performance [3, 5]. In-memory database systems [2, 8, 4] can execute transactions an order of magnitude faster than conventional database systems. While keeping the database in memory can greatly reduce read latencies, writes need to be made durable to ensure that the system can recover to a consistent state after a failure. Although inmemory database systems usually perform less write I/O due to the absence of disk-resident B-trees and heap files, high performance transaction processing generates a correspondingly high rate of update I/O. For restart recovery, disk-resident database systems read log records generated since the last checkpoint so as to recover to a consistent database state. On the other hand, in-memory database systems have to read the entire database as well as the log, and rebuild indices on database start-up. These steps may result in a long restart recovery time. Limiting the amount of data that must be read during recovery can be valuable for controlling the recovery time. In-memory database systems generally use one of two paradigms for persisting updates. One simple method is to write in-memory data to disk by using a direct mapping of an in-memory repre- Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. IMDM 15, August , Kohala Coast, HI, USA c 2015 ACM. ISBN /15/08... $15.00 DOI: sentation to persistent storage [4]. We refer to this method as as Update-In-Place (UIP). Physical data independence implies that the positions of updated objects in persistent storage do not necessarily match the order of the updates, leading to random writes. To address this problem, systems such as H-Store [5] and SiloR [9] checkpoint not only updated objects but also clean objects. This form of UIP using snapshotting (UIP-S) optimizes performance through the conversion of writes from random to sequential. Unlike UIP and UIP-S, in-memory database systems such as Hekaton [2] follow a different paradigm, under which all updates are appended to a series of checkpoint files on disk. This method, which we refer to as Copy-On-Write (COW), is advantageous in that it converts random I/O into sequential I/O. In addition, to alleviate the disruptive impact of checkpoints, some COW systems such as Hekaton spill out I/Os continuously. However, COW systems incur overhead due to the need to garbage collect outdated objects in checkpoint files. Systems such as RAMCloud [7] avoid this overhead by using copy-on-write both in memory and on persistent storage. Garbage collection can then be performed on the in-memory representation, and then persisted afterwards. We refer to this variant as COW-M, and to the original as COW-D. In this paper, we study these different persistent storage management techniques for in-memory database systems. Our goal is to characterize the I/O efficiency of these techniques, which we quantify as write amplification. Intuitively, write amplification is the ratio of the rate with which data are written to persistent storage to the rate at which the database is updated. Techniques with high write amplifications have poor I/O efficiency; They must be able to perform a lot of I/O to sustain a given database update rate. We formulate a simple analytical model for database I/O and use it to measure the write amplication factor for each storage management scheme shown in Table 1. We study the write amplification factor under both a random update model and a skewed update model. Our model also allows us to distinguish between random and sequential I/O, so that we can weigh random I/O more heavily than sequential I/O when calculating write amplification. Our model allows us to estimate to compare the write amplification factors of various storage management techniques, which is valuable for several reasons. First, the write amplification characteristics provide us with some insight into the different natures of update-in-place and copy-on-write storage managers, and to the settings in which each may be most appropriate. Second, improved I/O efficiency can contribute to lower cost for operating a database management system, since such a system must be configured with enough I/O capacity to sustain its anticipated performance. Third, in situations in which I/O capacity is constrained, improved I/O efficiency may lead to better system performance. It is also worth pointing out what our model does not do. Most
2 Storage Scheme Description Examples UIP persistent DB plus log, Shore-MT checkpoint dirty pages UIP-S persistent DB plus log, H-Store, SiloR checkpoint entire DB COW-D log-structured persistent Hekaton DB COW-M log-structured persistent RAMCloud and in-memory DB Table 1: Storage Management for In-Memory Database Systems importantly, our model does not attempt to directly predict the performance (e.g., transaction throughput or latency) of an in-memory DBMS. Such systems are complex and difficult to model in their full generality, and performance depends on many factors in addition to I/O efficiency. In particular, many systems try to hide as much I/O latency as possible by performing I/O operations outside of the critical transaction execution path. Thus, we expect that it would be extremely difficult to analytically model the effect of I/O on, say, transaction latency. Even if this were possible, the resulting model would be tied closely to a particular DBMS. Instead, we have developed a more abstract model that focuses on I/O efficiency, and that tries to capture only the key characteristics of copy-on-write and update-in-place storage managers, ignoring many implementation-specific details. The rest of this paper is organized as follows: Section 2 describes the systems from Table 1; Section 3 presents our write amplification analytical model for database I/O, and Section 4 presents the results of our analysis using this model. In Section 5, we generalize the model from Section 3 to account for update skew. Section 6 concludes the paper. 2. MANAGING PERSISTENT STORAGE Most OLTP DBMSs append updates to persistent logs, which provides durability without overwriting the persistent database. Since the storage capacity of the log is bounded, UIP and COW systems periodically or continuously clean their logs. This log cleaning operation in UIP systems is accomplished by checkpointing while in COW systems it is accomplished by log cleaning, a form of garbage collection. When done in batches (of updates), log cleaning is efficient though lazy log cleaning can result in longer restart recovery time due to the possibility of having a greater number of updates to process from the log. Thus, the frequency of log cleaning can impact normal system operation, I/O efficiency, and restart recovery time. This tradeoff between recovery time (due to log size) and I/O per update (IOPU) is present when providing in-memory DBMS persistence. Next, we describe how UIP and COW schemes (outlined in Table 1) manage their persistent storage. 2.1 Update-In-Place Scheme In UIP, persistent storage holds a copy of the in-memory database and the log to which committed updates are written. Restart recovery time is proportional to the size of the log, so its size is capped by checkpointing. During checkpointing, the DBMS copies updated objects from its in-memory database to its persistent database. In addition to this copy operation, the DBMS inserts a checkpoint record in the log to indicate the position where the last log records were checkpointed. After copying updated objects, the DBMS can delete its corresponding log records. In conventional disk-based DBMSs, the database consists of pages where each page contains multiple objects. Updated pages in memory are copied to persistent storage through checkpointing or when the buffer eviction policy is applied only some pages are buffered in memory because the size of memory space is limited. I/O resulting from checkpointing or buffer evictions may be random writes. In contrast, two representative UIP in-memory systems, H-Store [1] and SiloR [9], copy the entire in-memory database including both dirty and clean objects. We call this type of checkpoint a snapshot checkpoint under the Update-In-Place-Snapshot scheme (UIP-S). The advantage of UIP-S is in its simplicity and its ability to convert random writes to sequential database writes. The DBMS does not need to keep track of which objects are dirty, and checkpointing does not result in random writes. An additional benefit is that UIP-S can easily support the creation of multiple historical backups of the database, providing the ability to revert to a previous backup version in case of an unrecoverable failure. However, because a snapshot checkpoint copies both dirty and clean objects, the amount of checkpoint I/O is proportional to the database size, regardless of the update frequency. 2.2 Copy-On-Write Scheme In the copy-on-write scheme, persistent storage holds only a log, which is itself the database (this is sometimes referred to as logstructured storage). Like UIP, the log size needs to be bounded through log cleaning so as to prevent the log from growing indefinitely as the database is updated. In COW implementations, objects in the log are usually grouped together into a segment. When the log size reaches a set threshold, live objects from segments with the fewest live objects are re-inserted into the log while the rest are removed to reduce the total size of the log. Some COW-based systems may implement additional features. For example, Hekaton holds an additional shared log for both memory and disk DBMSs so as to work seamlessly with legacy diskbased DBMSs [2]. Hekaton checkpoints continuously by dumping updated objects to data files. In addition, Hekaton appends IDs of deleted (invalidated) objects to delta files. Each data file has a corresponding delta file that can be used to avoid loading deleted objects to the in-memory database during restart recovery. The specifics of log cleaning also vary among COW systems. Hekaton merges two adjacent data files and writes the merged data file into the same position from where the two data files were located. This is to ensure that segments are placed in epoch order in the system. In contrast, RAMCloud [7] merges multiple segments from arbitary locations, and inserts the merged segment into the head of the log as merged objects are treated the same as the latest objects updated by the application. RAMCloud maintains its in-memory COW representation also on disk (we call this the COW-M scheme). Instead of updating inplace in memory, RAMCloud appends updates without overwriting existing objects. Appending valid objects from old segments to new segments during log cleaning allows RAMCloud to avoid reading all of the old segments. A COW scheme would perform best when persistent storage is more than twice the size of the database in general, though the memory may not reach this size. To address this, RAMCloud compacts segments that contain the highest percentage of invalid objects, thereby avoiding immediate log cleaning on disk when memory space is limited. 3. WRITE AMPLIFICATION MODELS To allow us to compare the I/O efficiency of the persistent storage management schemes shown in Table 1, we have developed a simple analytic database I/O model that is general enough to capture all
3 parameter definition description D input database size input storage scale factor S D persistent storage capacity P input page size L S D persistent log capacity (UIP) P hot input update skew parameter D hot P hot D database hotspot size Table 2: Model Parameter Summary of the schemes. Our model predicts the write amplification that will occur under each of the storage management schemes. Intuitively, write amplification is the ratio of the I/O rate to the database update rate. For example, a write amplification of two means that the I/O to persistent storage will occur at twice the rate with which the (in-memory) database is updated. Write amplification is a measure of I/O inefficiency, i.e., lower values are better. A storage management scheme with low write amplification will place less demand on persistent storage than one with higher write efficiency. In general, we expect write amplification to depend on the amount of persistent storage space that is available. For this reason, our model parameterizes the amount of persistent storage, and all comparisons that we make among the storage management schemes are made on a constant-space basis. Specifically, we model the database as a set of objects, with cardinality D, and we assume that the capacity of persistent storage is D objects, where 1. Thus, controls how fully the database utilizes the persistent storage space. Larger values of result in lower utilization of persistent storage. Another reason for parameterizing persistent storage space in our model is that the amount of persistent storage is related to recovery time. Under all of the storage management schemes that we consider, it is necessary to read the contents of persistent storage to fully recover from a failure that results in the loss of the in-memory copy of the database. Thus, to the extent that I/O is the performance bottleneck during system recovery, D (the persistent storage capacity) can serve as an indirect measure of recovery time, and our constant space comparisons of storage management schemes are also constant recovery time comparisons. Our model assumes that the in-memory database is updated at a fixed rate of object updates per unit time, and that objects are selected randomly for update. For the purposes of the analyses in this section, we assume that all objects are equally likely to be updated. In Section 5, we generalize our analyses to account for update skew, i.e., settings in which some objects are more likely to be updated than others. Table 2 summarizes our model parameters, including and D. The remaining parameters will be introduced as we present our analyses. 3.1 Update-In-Place (UIP) To estimate the write expansion for UIP storage managers, we use a simple stop-and-copy checkpointing model. Under this model, updates are logged without checkpoint until the available log space is filled. Then the database temporarily halts update processing, performs a checkpoint, clears the persistent log, and then resumes update processing. Let L = D D = ( 1)D represent the amount of log space in persistent storage. Under the stopand-copy model, the DBMS must checkpoint at least once every L updates since each update consumes one unit of log space. The UIP storage manager divides the database into pages of size P objects. Assuming a uniform distribution of updates over pages, the probability of a database page receiving at least one update during a checkpoint interval, which we refer to as P dirty, is given by ( ) D P L ( ) D P ( 1)D P dirty = 1 = 1 (1) D D Thus, the expected amount of data that will need to be written to the persistent database during each checkpoint interval is P dirty D. In addition, the DBMS will write L objects to the log during the interval. The write amplification factor for UIP, which we refer to as IOPU UIP (I/O Per Update) is the ratio of the number of objects written to persistent storage during each interval to the number of updates to the in-memory database: IOPU UIP = P dirtyd + L L = P dirty (2) The UIP-S storage manager is similar to UIP, except that it writes the entire database to persistent storage during each checkpoint. Thus, the write amplification factor for UIP-S is IOPU = D + L L = Accounting for Random I/O An important property of UIP storage managers is that I/O for database checkpoints is random. In contrast, all I/O generated by UIP-S and COW storage managers is sequential. Indeed, this is one of the key factors that has motivated COW storage management. Thus, we would like to adjust the UIP model to account for random I/O. To do this, we apply a penalty factor ρ 1 to the I/O cost of UIP checkpoints. Intuitively, if the I/O operations performed by the UIP checkpoint are, for example, 3 times as expensive as sequential I/O operations, then we want ρ to be three. We expect the relative cost of a UIP checkpoint s I/O operations to (ρ) to depend on two factors. The first is P, the size of the database pages, since writing larger pages may be more efficient than writing smaller pages. The second is P dirty, which determines the number of pages written during each checkpoint. Because of I/O scheduling, I/O performed by checkpoints that write many pages can be more efficient than I/O performed by checkpoints that write fewer pages. Thus, we define ρ as follows: (3) ρ = P dirty + (1 P dirty ) Penalty(P) (4) where Penalty(P) is a device-specific function that characterizes the cost of random writes of pages of size P, relative to the cost of sequential writes of pages of the same size. Thus, as P dirty approaches one, and each checkpoint writes more pages, ρ will approach one. Conversely, as P dirty approaches zero, and the checkpoint writes few pages, ρ will approach Penalty(P). With the introduction of ρ, the revised model of the write amplification of UIP becomes IOPU UIP = ρp dirtyd + L L = ρp dirty (5) We can estimate Penalty(P) for specific types of I/O devices using simple calibration tests which measure sequential and random I/O bandwidth at different page sizes. We then take Penalty(P) to be the ratio of measured sequential write bandwidth to measured random write bandwidth for pages of size P. For example, Figure 1 shows the Penalty(P) determined using this method for two devices, an Intel SSDSC2BA200G 200GB SATA SSD, and a Seagate ST1000NM0033 1TB 7200RPM SATA disk drive. For the
4 Random Write Penalty W The Number of Tuples Per Page (P) Figure 1: Random Write Penalty for Two Devices Storage S Figure 2: Copy-on-write model Pvalid R SSD HDD UIP results presented in the rest of this paper, we have used these two penalty functions to model UIP write amplification on SSDs and disk drives (HDDs). 3.2 Copy-On-Write (COW) COW storage managers use all of persistent storage as a log. We begin by modeling the behavior of COW-D, and then show how this model can be revised to model COW-M. To estimate write amplification for COW-M, we model its use of persistent storage as shown in Figure 2. refers to the rate of database updates in the in-memory database. As these updates occur, the storage manager appends a log record for each update to one end of the log - the left end in Figure 2. These log insertions occur at rate. To ensure that the log does not overflow persistent storage, the storage manager removes entries from the other end of the log, at rate R. It checks each removed entry to determine whether the object version that it describes is still valid, or is garbage. If the version is still valid, the removed log record for that version is re-inserted at the head of the log. Otherwise it is discarded. We use P valid to represent the probability that a removed object version is still valid, and W to refer to the re-insertion rate, i.e., the rate with which valid versions are re-inserted, as shown in Figure 2. With these definitions, we can define the write amplification factor for COW storage managers, IOPU COW D, as: IOPU COW D = + R + W Note that, unlike UIP storage managers, the I/O for a COW-D storage manager includes reads as well as writes, since the log must be read for garbage collection. We include both reads and writes to persistent storage in the write amplification factor. (6) In steady state, it must be the case that R = + W. IOPU COW D = 2( + W ) From our definitions, W depends on and P valid. Thus, we can re-write Equation 7 as IOPU COW D = (7) 2 1 P valid (8) To estimate P valid, consider an arbitrary log entry inserted into the log by the storage manager. After D 1 subsequent insertions, that log entry will be at the tail of the log, about to be cleaned. Let U represent the number of those subsequent insertions that are not re-insertions. We can write P valid = W = W D 1 U = R + W D 1 Since only one valid version of each database record exists in the log at any time, the maximum number of re-insertions that can occur before our arbitrary log entry reaches the end of the log is D 1. Thus, we also know that ( 1)D U D 1. For arbitrary update distributions, these lower and upper bounds on U give corresponding upper and lower bounds on IOPU COW D. For the special case in which updates are uniformly distributed over the database objects, we can also relate P valid to U by the following expression. ( ) D 1 U P valid = (10) D This represents the probability that a newly logged update is not invalidated before it reaches the tail of the log. Using our general bounds on U, we can solve Equation 10 numerically for P valid, and hence obtain IOPU COW D from Equation Modeling COW-M The behavior of COW-M is similar to that of COW-D, except that it does not need to read the log from persistent storage to determine which log entries require re-insertion. Thus, we can write: IOPU = + W (11) Following the same arguments used for the COW-D model, we can write: 1 IOPU = (12) 1 P valid To calculate IOPU for the special case of a uniform update distribution, we calculate P valid using Equation 10, and use this value to determine IOPU from Equation EVALUATION Before comparing UIP and COW storage managers, we begin with an analysis of the effect of page size on UIP. Figure 3 shows the write expansion factor (IOPU) for UIP storage managers as functions of and the page size, P. Figure 3(b) shows the same data as Figure 3(a), but focused on a range of small values of. For Figure 3, we have assumed that persistent storage is implemented with hard disks, and the write expansion factors have been adjusted with a random I/O penalty as described in Section In this and all other figures presented in this paper, we have fixed the database size, D, at 10 6 objects. The message from this figure is clear. Unless the log is very small, UIP-S is the preferred update-in-place storage manager. Recall that the available log size under our model is ( 1)D. Thus, (9)
5 UIP(8K) UIP(1K) UIP(10) UIP(1) UIP(8K) UIP(1K) UIP(10) UIP(1) (a) Large UIP(8K) UIP(1K) UIP(10) UIP(1) 20 Figure 4: UIP: Uniform Random Updates, SSD UIP(10) HDD UIP(10) SSD COW D (b) Small Figure 3: UIP: Uniform Random Updates, HDD Figure 5: UIP vs. COW, Uniform Random Updates small values of correspond to storage managers configured to use little log space typically to allow fast recovery from failures. All storage managers have high write amplifications in such small- configurations, but UIP may be not as bad as UIP-S for extremely small. Figure 4 shows the same comparison, but under the assumption that persistent storage is implemented with SSDs. Here, UIP (with small pages) beats UIP-S for a wider range of (small) al pha values, since the penalty for non-sequential updates is smaller on SSDs. However, as was the case with HDDs, larger values of drive write amplification down quickly because checkpointing can occur less frequently, and UIP quickly converges to UIP-S under those conditions. 4.1 UIP vs. COW Figure 5 shows the IOPU of the two COW storage managers, as well as several of the UIP managers. COW-M has lower write amplification than COW-D, and converges to the ideal write amplification factor of 1 as grows. The important distinction between COW-M and COW-D that accounts for this is that COW-D must read from persistent storage to perform cleaning, while COW-M does not. Thus, for large, IOPU COW D converges to 2, rather than 1, since every object update will result in one object being written (to commit the update) and one being read (to clean the committed update from the log). Of course, the disadvantage of COW-M is that it requires a log-structured in-memory data representation that mirrors the representation in persistent storage. For all, UIP-S write-amplification is either comparable to, or slightly better than, that of COW-D. Both types of storage managers perform only sequential I/O. However, on an equal-space basis, UIP checkpointing is a more I/O-efficient way to limit log size than COW-D s log cleaning. For very large values of, IOPU converges to the ideal value of 1. Finally, we note that in situations where fast recovery is critical (small ), UIP(10) is competitive with even COW-M, provided that there is not a large penalty for random I/O. 5. ACCOUNTING FOR SKEW In Section 3, we presented write amplification models that were developed under an assumption of uniformly distributed database updates. In this section, we generalize those models to account for skewed updates. We use a Pareto skew model, in which P hot % of the database receives (100 P hot )% of the updates. We vary the
6 parameter P hot to control the skew. 5.1 Update-In-Place (UIP) The write amplification model for the update-in-place storage managers is the same with skew as without, except for the calculation of P dirty. With the introduction of skew, we assume that hot objects and cold objects are clustered, resulting in hot pages and cold pages. The probability that a page is dirty will depend on whether it is hot or cold. With skew, we replace our previous calculation of P dirty (Equation 1) with the skew-aware calculation given by Equation 13. P dirty = P dirty hot P hot + P dirty cold (1 P hot ) ( ) Dhot P (1 Phot )L P dirty hot = 1 D hot ( ) (D Dhot ) P Phot L P dirty cold = 1 (D D hot ) (13) 5.2 Copy-On-Write (COW) Incorporating skew into the copy-on-write model is more challenging. The write amplification for copy-on-write storage managers depends on P valid, the probability that determines the log reinsertion rate. Equation 9 for P valid holds for all update distributions. However, Equation 10 does not. With skewed updates, cold objects are more likely to be re-inserted than hot objects since they are less likely to be over-written. Thus, P valid is a complex function of P hot and the log size. Rather than solving analytically for IOPU COW D and IOPU when updates are skewed, we instead use discrete-event simulations of the COW model (Figure 2) to estimate both write amplification factors. 5.3 Evaluation of Skew s Impact Figures 6 compares the I/O per update (IOPU) for both UIP and COW storage managers at several different levels of skew. The UIP-S storage manager write amplification is insensitive to skew, since it always checkpoints the entire database. However, the write amplification of the other UIP storage managers is very sensitive. Higher skew is beneficial to UIP storage managers because it reduces the amount of data that must be written during each checkpoint. When is small, UIP to SSDs is a winning strategy. In contrast, skew has a mildly negative effect on the COW storage managers. Cold objects have relatively long lifetimes, and thus are more likely to have to be salvaged and re-inserted during log cleaning, increasing the I/O cost of the cleaner. Some techniques have been proposed to address this issue [6], but they are not included in our analysis. 6. CONCLUSION In this paper, we have presented a write amplification model for in-memory database storage managers. Our model allows us to quantify and compare the I/O efficiency of two broad classes of storage managers: those that use update-in-place, and those that use copy-on-write. Our model is parameterized by space efficiency, which makes it easy for us to consider the write amplification of different storage managers on a constant-space basis. Among the storage managers we tested, COW-M was the most I/O efficient (lowest write amplification) in many settings. However, unlike the other storage managers, COW-M requires a log structure in memory that mirrors the log structure in persistent storage, which may be a significant limitation. For example, COW-M UIP(10) HDD UIP(10) SSD COW D (a) Light Skew (60-40) UIP(10) HDD UIP(10) SSD COW D (b) Medium Skew (80-20) UIP(10) HDD UIP(10) SSD COW D (c) High Skew (99-1) Figure 6: Write Amplification Under Skew Distribution
7 may preclude the storage manager from using a cache-friendly inmemory data organization that is optimized for efficient reads. Among the storage managers that do not constrain the in-memory organization of the database, UIP-S is an appealing option in many settings. It is always at least as I/O efficient as COW-D, and in most situations is also at least as good as page-based UIP alternatives. UIP-S is also insensitive to update skew, and it requires only sequential writes to persistent storage. Page-level UIP schemes, which are commonly used in disk-based database systems, make sense for in-memory databases only in very specific situations. For example, they can be effective when the capacity of persistent storage (or recovery time) is tightly constrained, provided that persistent storage can support random writes efficiently. They may also be effective if database updates are highly skewed. 7. REFERENCES [1] H-Store Manual hstore.cs.brown.edu/ documentation/. [2] C. Diaconu, C. Freedman, E. Ismert, P.-A. Larson, P. Mittal, R. Stonecipher, N. Verma, and M. Zwilling. Hekaton: SQL server s memory-optimized OLTP engine. In Proc. ACM SIGMOD, pages , New York, New York, USA, June ACM. [3] S. Harizopoulos, D. J. Abadi, S. Madden, and M. Stonebraker. Oltp through the looking glass, and what we found there. In Proc. ACM SIGMOD, pages , New York, NY, USA, ACM. [4] R. Johnson, I. Pandis, N. Hardavellas, A. Ailamaki, and B. Falsafi. Shore-MT: A Scalable Storage Manager for the Multicore Era. In Proc. EDBT, pages 24 35, New York, NY, USA, ACM. [5] R. Kallman, H. Kimura, J. Natkins, A. Pavlo, A. Rasin, S. Zdonik, E. P. C. Jones, S. Madden, M. Stonebraker, Y. Zhang, J. Hugg, and D. J. Abadi. H-Store: a high-performance, distributed main memory transaction processing system. The VLDB Journal, 1(2): , [6] M. Rosenblum and J. K. Ousterhout. The design and implementation of a log-structured file system. ACM SIGOPS OSR, [7] S. M. Rumble, A. Kejriwal, and J. Ousterhout. Log-structured memory for DRAM-based storage. In Proc. USENIX FAST. USENIX Association, Feb [8] S. Tu, W. Zheng, E. Kohler, B. Liskov, and S. Madden. Speedy transactions in multicore in-memory databases. In Proc. SOSP. USENIX Association, Nov [9] W. Zheng, S. Tu, E. Kohler, and B. Liskov. Fast Databases with Fast Durability and Recovery Through Multicore Parallelism. In Proc. SOSP, pages USENIX Association, 2014.
Lab Evaluation of NetApp Hybrid Array with Flash Pool Technology
Lab Evaluation of NetApp Hybrid Array with Flash Pool Technology Evaluation report prepared under contract with NetApp Introduction As flash storage options proliferate and become accepted in the enterprise,
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE
WITH A FUSION POWERED SQL SERVER 2014 IN-MEMORY OLTP DATABASE 1 W W W. F U S I ON I O.COM Table of Contents Table of Contents... 2 Executive Summary... 3 Introduction: In-Memory Meets iomemory... 4 What
Seeking Fast, Durable Data Management: A Database System and Persistent Storage Benchmark
Seeking Fast, Durable Data Management: A Database System and Persistent Storage Benchmark In-memory database systems (IMDSs) eliminate much of the performance latency associated with traditional on-disk
Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database
WHITE PAPER Improve Business Productivity and User Experience with a SanDisk Powered SQL Server 2014 In-Memory OLTP Database 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Executive
The Classical Architecture. Storage 1 / 36
1 / 36 The Problem Application Data? Filesystem Logical Drive Physical Drive 2 / 36 Requirements There are different classes of requirements: Data Independence application is shielded from physical storage
Hybrid Storage Management for Database Systems
Hybrid Storage Management for Database Systems Xin Liu University of Waterloo, Canada [email protected] Kenneth Salem University of Waterloo, Canada [email protected] ABSTRACT The use of flash-based
EMC XtremSF: Delivering Next Generation Performance for Oracle Database
White Paper EMC XtremSF: Delivering Next Generation Performance for Oracle Database Abstract This white paper addresses the challenges currently facing business executives to store and process the growing
Compilation in the Microsoft SQL Server Hekaton Engine
Compilation in the Microsoft SQL Server Hekaton Engine Craig Freedman [email protected] Erik Ismert [email protected] Per-Ake Larson [email protected] Abstract Hekaton is a new database engine
Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card
Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card Version 1.0 April 2011 DB15-000761-00 Revision History Version and Date Version 1.0, April 2011 Initial
Flash for Databases. September 22, 2015 Peter Zaitsev Percona
Flash for Databases September 22, 2015 Peter Zaitsev Percona In this Presentation Flash technology overview Review some of the available technology What does this mean for databases? Specific opportunities
Energy aware RAID Configuration for Large Storage Systems
Energy aware RAID Configuration for Large Storage Systems Norifumi Nishikawa [email protected] Miyuki Nakano [email protected] Masaru Kitsuregawa [email protected] Abstract
89 Fifth Avenue, 7th Floor. New York, NY 10003. www.theedison.com 212.367.7400. White Paper. HP 3PAR Adaptive Flash Cache: A Competitive Comparison
89 Fifth Avenue, 7th Floor New York, NY 10003 www.theedison.com 212.367.7400 White Paper HP 3PAR Adaptive Flash Cache: A Competitive Comparison Printed in the United States of America Copyright 2014 Edison
The Data Placement Challenge
The Data Placement Challenge Entire Dataset Applications Active Data Lowest $/IOP Highest throughput Lowest latency 10-20% Right Place Right Cost Right Time 100% 2 2 What s Driving the AST Discussion?
Automated Data-Aware Tiering
Automated Data-Aware Tiering White Paper Drobo s revolutionary new breakthrough technology automates the provisioning, deployment, and performance acceleration for a fast tier of SSD storage in the Drobo
Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments
Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments Applied Technology Abstract This white paper introduces EMC s latest groundbreaking technologies,
In-Memory Performance for Big Data
In-Memory Performance for Big Data Goetz Graefe, Haris Volos, Hideaki Kimura, Harumi Kuno, Joseph Tucek, Mark Lillibridge, Alistair Veitch VLDB 2014, presented by Nick R. Katsipoulakis A Preliminary Experiment
An Approach to High-Performance Scalable Temporal Object Storage
An Approach to High-Performance Scalable Temporal Object Storage Kjetil Nørvåg Department of Computer and Information Science Norwegian University of Science and Technology 791 Trondheim, Norway email:
How To Store Data On An Ocora Nosql Database On A Flash Memory Device On A Microsoft Flash Memory 2 (Iomemory)
WHITE PAPER Oracle NoSQL Database and SanDisk Offer Cost-Effective Extreme Performance for Big Data 951 SanDisk Drive, Milpitas, CA 95035 www.sandisk.com Table of Contents Abstract... 3 What Is Big Data?...
CS 153 Design of Operating Systems Spring 2015
CS 153 Design of Operating Systems Spring 2015 Lecture 22: File system optimizations Physical Disk Structure Disk components Platters Surfaces Tracks Arm Track Sector Surface Sectors Cylinders Arm Heads
IS IN-MEMORY COMPUTING MAKING THE MOVE TO PRIME TIME?
IS IN-MEMORY COMPUTING MAKING THE MOVE TO PRIME TIME? EMC and Intel work with multiple in-memory solutions to make your databases fly Thanks to cheaper random access memory (RAM) and improved technology,
Accelerating Server Storage Performance on Lenovo ThinkServer
Accelerating Server Storage Performance on Lenovo ThinkServer Lenovo Enterprise Product Group April 214 Copyright Lenovo 214 LENOVO PROVIDES THIS PUBLICATION AS IS WITHOUT WARRANTY OF ANY KIND, EITHER
IN-MEMORY DATABASE SYSTEMS. Prof. Dr. Uta Störl Big Data Technologies: In-Memory DBMS - SoSe 2015 1
IN-MEMORY DATABASE SYSTEMS Prof. Dr. Uta Störl Big Data Technologies: In-Memory DBMS - SoSe 2015 1 Analytical Processing Today Separation of OLTP and OLAP Motivation Online Transaction Processing (OLTP)
Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software
WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications
EMC VFCACHE ACCELERATES ORACLE
White Paper EMC VFCACHE ACCELERATES ORACLE VFCache extends Flash to the server FAST Suite automates storage placement in the array VNX protects data EMC Solutions Group Abstract This white paper describes
EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server
White Paper EMC XtremSF: Delivering Next Generation Storage Performance for SQL Server Abstract This white paper addresses the challenges currently facing business executives to store and process the growing
COS 318: Operating Systems
COS 318: Operating Systems File Performance and Reliability Andy Bavier Computer Science Department Princeton University http://www.cs.princeton.edu/courses/archive/fall10/cos318/ Topics File buffer cache
Recovery Protocols For Flash File Systems
Recovery Protocols For Flash File Systems Ravi Tandon and Gautam Barua Indian Institute of Technology Guwahati, Department of Computer Science and Engineering, Guwahati - 781039, Assam, India {r.tandon}@alumni.iitg.ernet.in
Fast Crash Recovery in RAMCloud
Fast Crash Recovery in RAMCloud Diego Ongaro, Stephen M. Rumble, Ryan Stutsman, John Ousterhout, and Mendel Rosenblum Stanford University ABSTRACT RAMCloud is a DRAM-based storage system that provides
Virtuoso and Database Scalability
Virtuoso and Database Scalability By Orri Erling Table of Contents Abstract Metrics Results Transaction Throughput Initializing 40 warehouses Serial Read Test Conditions Analysis Working Set Effect of
CASE STUDY: Oracle TimesTen In-Memory Database and Shared Disk HA Implementation at Instance level. -ORACLE TIMESTEN 11gR1
CASE STUDY: Oracle TimesTen In-Memory Database and Shared Disk HA Implementation at Instance level -ORACLE TIMESTEN 11gR1 CASE STUDY Oracle TimesTen In-Memory Database and Shared Disk HA Implementation
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
Flash Controller Architecture for All Flash Arrays
Flash Controller Architecture for All Flash Arrays Andy Walls Distinguished Engineer, IBM Systems and Technology Group CTO, Flash Systems and Technology Santa Clara, CA August 2013 1 Once upon a Time,.....
Remote Copy Technology of ETERNUS6000 and ETERNUS3000 Disk Arrays
Remote Copy Technology of ETERNUS6000 and ETERNUS3000 Disk Arrays V Tsutomu Akasaka (Manuscript received July 5, 2005) This paper gives an overview of a storage-system remote copy function and the implementation
Fast Crash Recovery in RAMCloud
Fast Crash Recovery in RAMCloud Diego Ongaro, Stephen M. Rumble, Ryan Stutsman, John Ousterhout, and Mendel Rosenblum Stanford University ABSTRACT RAMCloud is a DRAM-based storage system that provides
Benchmarking Hadoop & HBase on Violin
Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages
Using Synology SSD Technology to Enhance System Performance. Based on DSM 5.2
Using Synology SSD Technology to Enhance System Performance Based on DSM 5.2 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD Cache as Solution...
FAWN - a Fast Array of Wimpy Nodes
University of Warsaw January 12, 2011 Outline Introduction 1 Introduction 2 3 4 5 Key issues Introduction Growing CPU vs. I/O gap Contemporary systems must serve millions of users Electricity consumed
Speeding Up Cloud/Server Applications Using Flash Memory
Speeding Up Cloud/Server Applications Using Flash Memory Sudipta Sengupta Microsoft Research, Redmond, WA, USA Contains work that is joint with B. Debnath (Univ. of Minnesota) and J. Li (Microsoft Research,
Using Synology SSD Technology to Enhance System Performance Synology Inc.
Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...
Outline. Failure Types
Outline Database Management and Tuning Johann Gamper Free University of Bozen-Bolzano Faculty of Computer Science IDSE Unit 11 1 2 Conclusion Acknowledgements: The slides are provided by Nikolaus Augsten
PrimaryIO Application Performance Acceleration Date: July 2015 Author: Tony Palmer, Senior Lab Analyst
ESG Lab Spotlight PrimaryIO Application Performance Acceleration Date: July 215 Author: Tony Palmer, Senior Lab Analyst Abstract: PrimaryIO Application Performance Acceleration (APA) is designed to provide
Integrating Flash-based SSDs into the Storage Stack
Integrating Flash-based SSDs into the Storage Stack Raja Appuswamy, David C. van Moolenbroek, Andrew S. Tanenbaum Vrije Universiteit, Amsterdam April 19, 2012 Introduction: Hardware Landscape $/GB of flash
OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006
OPTIMIZING EXCHANGE SERVER IN A TIERED STORAGE ENVIRONMENT WHITE PAPER NOVEMBER 2006 EXECUTIVE SUMMARY Microsoft Exchange Server is a disk-intensive application that requires high speed storage to deliver
SAP HANA - Main Memory Technology: A Challenge for Development of Business Applications. Jürgen Primsch, SAP AG July 2011
SAP HANA - Main Memory Technology: A Challenge for Development of Business Applications Jürgen Primsch, SAP AG July 2011 Why In-Memory? Information at the Speed of Thought Imagine access to business data,
PIONEER RESEARCH & DEVELOPMENT GROUP
SURVEY ON RAID Aishwarya Airen 1, Aarsh Pandit 2, Anshul Sogani 3 1,2,3 A.I.T.R, Indore. Abstract RAID stands for Redundant Array of Independent Disk that is a concept which provides an efficient way for
Flash-Friendly File System (F2FS)
Flash-Friendly File System (F2FS) Feb 22, 2013 Joo-Young Hwang ([email protected]) S/W Dev. Team, Memory Business, Samsung Electronics Co., Ltd. Agenda Introduction FTL Device Characteristics
Flash Memory Arrays Enabling the Virtualized Data Center. July 2010
Flash Memory Arrays Enabling the Virtualized Data Center July 2010 2 Flash Memory Arrays Enabling the Virtualized Data Center This White Paper describes a new product category, the flash Memory Array,
CSE 120 Principles of Operating Systems
CSE 120 Principles of Operating Systems Fall 2004 Lecture 13: FFS, LFS, RAID Geoffrey M. Voelker Overview We ve looked at disks and file systems generically Now we re going to look at some example file
HP Smart Array Controllers and basic RAID performance factors
Technical white paper HP Smart Array Controllers and basic RAID performance factors Technology brief Table of contents Abstract 2 Benefits of drive arrays 2 Factors that affect performance 2 HP Smart Array
An Oracle White Paper November 2010. Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager
An Oracle White Paper November 2010 Backup and Recovery with Oracle s Sun ZFS Storage Appliances and Oracle Recovery Manager Introduction...2 Oracle Backup and Recovery Solution Overview...3 Oracle Recovery
SQL Server Transaction Log from A to Z
Media Partners SQL Server Transaction Log from A to Z Paweł Potasiński Product Manager Data Insights [email protected] http://blogs.technet.com/b/sqlblog_pl/ Why About Transaction Log (Again)? http://zine.net.pl/blogs/sqlgeek/archive/2008/07/25/pl-m-j-log-jest-za-du-y.aspx
In-Memory Columnar Databases HyPer. Arto Kärki University of Helsinki 30.11.2012
In-Memory Columnar Databases HyPer Arto Kärki University of Helsinki 30.11.2012 1 Introduction Columnar Databases Design Choices Data Clustering and Compression Conclusion 2 Introduction The relational
The Design and Implementation of a Log-Structured File System
The Design and Implementation of a Log-Structured File System Mendel Rosenblum and John K. Ousterhout Electrical Engineering and Computer Sciences, Computer Science Division University of California Berkeley,
Big Data Functionality for Oracle 11 / 12 Using High Density Computing and Memory Centric DataBase (MCDB) Frequently Asked Questions
Big Data Functionality for Oracle 11 / 12 Using High Density Computing and Memory Centric DataBase (MCDB) Frequently Asked Questions Overview: SGI and FedCentric Technologies LLC are pleased to announce
Fast Checkpoint and Recovery Techniques for an In-Memory Database. Wenting Zheng
Fast Checkpoint and Recovery Techniques for an In-Memory Database by Wenting Zheng S.B., Massachusetts Institute of Technology (2013) Submitted to the Department of Electrical Engineering and Computer
LEVERAGING FLASH MEMORY in ENTERPRISE STORAGE. Matt Kixmoeller, Pure Storage
LEVERAGING FLASH MEMORY in ENTERPRISE STORAGE Matt Kixmoeller, Pure Storage SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted. Member companies
Hekaton: SQL Server s Memory-Optimized OLTP Engine
Hekaton: SQL Server s Memory-Optimized OLTP Engine Cristian Diaconu, Craig Freedman, Erik Ismert, Per-Åke Larson, Pravin Mittal, Ryan Stonecipher, Nitin Verma, Mike Zwilling Microsoft {cdiaconu, craigfr,
The Use of Flash in Large-Scale Storage Systems. [email protected]
The Use of Flash in Large-Scale Storage Systems [email protected] 1 Seagate s Flash! Seagate acquired LSI s Flash Components division May 2014 Selling multiple formats / capacities today Nytro
Maximizing SQL Server Virtualization Performance
Maximizing SQL Server Virtualization Performance Michael Otey Senior Technical Director Windows IT Pro SQL Server Pro 1 What this presentation covers Host configuration guidelines CPU, RAM, networking
Optimizing SQL Server AlwaysOn Implementations with OCZ s ZD-XL SQL Accelerator
White Paper Optimizing SQL Server AlwaysOn Implementations with OCZ s ZD-XL SQL Accelerator Delivering Accelerated Application Performance, Microsoft AlwaysOn High Availability and Fast Data Replication
PARALLELS CLOUD STORAGE
PARALLELS CLOUD STORAGE Performance Benchmark Results 1 Table of Contents Executive Summary... Error! Bookmark not defined. Architecture Overview... 3 Key Features... 5 No Special Hardware Requirements...
Flash at the price of disk Redefining the Economics of Storage. Kris Van Haverbeke Enterprise Marketing Manager Dell Belux
Flash at the price of disk Redefining the Economics of Storage Kris Van Haverbeke Enterprise Marketing Manager Dell Belux We believe that Redefining the Economics of Storage helps businesses achieve their
Best Practices for Architecting Storage in Virtualized Environments
Best Practices for Architecting Storage in Virtualized Environments Leverage Advances in Storage Technology to Accelerate Performance, Simplify Management, and Save Money in Your Virtual Server Environment
Preview of Oracle Database 12c In-Memory Option. Copyright 2013, Oracle and/or its affiliates. All rights reserved.
Preview of Oracle Database 12c In-Memory Option 1 The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any
3PAR Fast RAID: High Performance Without Compromise
3PAR Fast RAID: High Performance Without Compromise Karl L. Swartz Document Abstract: 3PAR Fast RAID allows the 3PAR InServ Storage Server to deliver higher performance with less hardware, reducing storage
How To Scale Myroster With Flash Memory From Hgst On A Flash Flash Flash Memory On A Slave Server
White Paper October 2014 Scaling MySQL Deployments Using HGST FlashMAX PCIe SSDs An HGST and Percona Collaborative Whitepaper Table of Contents Introduction The Challenge Read Workload Scaling...1 Write
Storage Architecture in XProtect Corporate
Milestone Systems White Paper Storage Architecture in XProtect Corporate Prepared by: John Rasmussen Product Manager XProtect Corporate Milestone Systems Date: 7 February, 2011 Page 1 of 20 Table of Contents
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
The Bw-Tree Key-Value Store and Its Applications to Server/Cloud Data Management in Production
The Bw-Tree Key-Value Store and Its Applications to Server/Cloud Data Management in Production Sudipta Sengupta Joint work with Justin Levandoski and David Lomet (Microsoft Research) And Microsoft Product
Using Synology SSD Technology to Enhance System Performance Synology Inc.
Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_WP_ 20121112 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD
Physical Data Organization
Physical Data Organization Database design using logical model of the database - appropriate level for users to focus on - user independence from implementation details Performance - other major factor
TECHNICAL OVERVIEW HIGH PERFORMANCE, SCALE-OUT RDBMS FOR FAST DATA APPS RE- QUIRING REAL-TIME ANALYTICS WITH TRANSACTIONS.
HIGH PERFORMANCE, SCALE-OUT RDBMS FOR FAST DATA APPS RE- QUIRING REAL-TIME ANALYTICS WITH TRANSACTIONS Overview VoltDB is a fast in-memory relational database system (RDBMS) for high-throughput, operational
Storage in Database Systems. CMPSCI 445 Fall 2010
Storage in Database Systems CMPSCI 445 Fall 2010 1 Storage Topics Architecture and Overview Disks Buffer management Files of records 2 DBMS Architecture Query Parser Query Rewriter Query Optimizer Query
SQL Server 2014 New Features/In- Memory Store. Juergen Thomas Microsoft Corporation
SQL Server 2014 New Features/In- Memory Store Juergen Thomas Microsoft Corporation AGENDA 1. SQL Server 2014 what and when 2. SQL Server 2014 In-Memory 3. SQL Server 2014 in IaaS scenarios 2 SQL Server
OLTP Meets Bigdata, Challenges, Options, and Future Saibabu Devabhaktuni
OLTP Meets Bigdata, Challenges, Options, and Future Saibabu Devabhaktuni Agenda Database trends for the past 10 years Era of Big Data and Cloud Challenges and Options Upcoming database trends Q&A Scope
RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University
RAMCloud and the Low- Latency Datacenter John Ousterhout Stanford University Most important driver for innovation in computer systems: Rise of the datacenter Phase 1: large scale Phase 2: low latency Introduction
Analyzing Big Data with Splunk A Cost Effective Storage Architecture and Solution
Analyzing Big Data with Splunk A Cost Effective Storage Architecture and Solution Jonathan Halstuch, COO, RackTop Systems [email protected] Big Data Invasion We hear so much on Big Data and
Anti-Caching: A New Approach to Database Management System Architecture
Anti-Caching: A New Approach to Database Management System Architecture Justin DeBrabant Andrew Pavlo Stephen Tu Brown University Brown University MIT CSAIL [email protected] [email protected] [email protected]
Enabling Efficient OS Paging for Main-Memory OLTP Databases
Enabling Efficient OS Paging for Main-Memory OLTP Databases Radu Stoica École Polytechnique Fédérale de Lausanne [email protected] Anastasia Ailamaki École Polytechnique Fédérale de Lausanne [email protected]
WHITE PAPER. Drobo TM Hybrid Storage TM
WHITE PAPER Drobo TM Hybrid Storage TM Table of Contents Introduction...3 What is Hybrid Storage?...4 SSDs Enable Hybrid Storage...4 One Pool, Multiple Tiers...5 Fully Automated Tiering...5 Tiering Without
Colgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP
selects SAP HANA to improve the speed of business analytics with IBM and SAP Founded in 1806, is a global consumer products company which sells nearly $17 billion annually in personal care, home care,
An Evaluation of Strict Timestamp Ordering Concurrency Control for Main-Memory Database Systems
An Evaluation of Strict Timestamp Ordering Concurrency Control for Main-Memory Database Systems Stephan Wolf, Henrik Mühe, Alfons Kemper, Thomas Neumann Technische Universität München {wolfst,muehe,kemper,neumann}@in.tum.de
DataStax Enterprise, powered by Apache Cassandra (TM)
PerfAccel (TM) Performance Benchmark on Amazon: DataStax Enterprise, powered by Apache Cassandra (TM) Disclaimer: All of the documentation provided in this document, is copyright Datagres Technologies
Buffer-Aware Garbage Collection for NAND Flash Memory-Based Storage Systems
Buffer-Aware Garbage Collection for NAND Flash Memory-Based Storage Systems Sungin Lee, Dongkun Shin and Jihong Kim School of Computer Science and Engineering, Seoul National University, Seoul, Korea {chamdoo,
In-Memory Data Management for Enterprise Applications
In-Memory Data Management for Enterprise Applications Jens Krueger Senior Researcher and Chair Representative Research Group of Prof. Hasso Plattner Hasso Plattner Institute for Software Engineering University
p-oftl: An Object-based Semantic-aware Parallel Flash Translation Layer
p-oftl: An Object-based Semantic-aware Parallel Flash Translation Layer Wei Wang, Youyou Lu, and Jiwu Shu Department of Computer Science and Technology, Tsinghua University, Beijing, China Tsinghua National
Evaluation Report: Supporting Microsoft Exchange on the Lenovo S3200 Hybrid Array
Evaluation Report: Supporting Microsoft Exchange on the Lenovo S3200 Hybrid Array Evaluation report prepared under contract with Lenovo Executive Summary Love it or hate it, businesses rely on email. It
Everything you need to know about flash storage performance
Everything you need to know about flash storage performance The unique characteristics of flash make performance validation testing immensely challenging and critically important; follow these best practices
LogBase: A Scalable Log structured Database System in the Cloud
: A Scalable Log structured Database System in the Cloud Hoang Tam Vo #1, Sheng Wang #2, Divyakant Agrawal 3, Gang Chen 4, Beng Chin Ooi # # National University of Singapore, University of California,
RAID Basics Training Guide
RAID Basics Training Guide Discover a Higher Level of Performance RAID matters. Rely on Intel RAID. Table of Contents 1. What is RAID? 2. RAID Levels RAID 0 RAID 1 RAID 5 RAID 6 RAID 10 RAID 0+1 RAID 1E
Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS
Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS M ost storage vendors now offer all-flash storage arrays, and many modern organizations recognize the need for these highperformance
New Features in PSP2 for SANsymphony -V10 Software-defined Storage Platform and DataCore Virtual SAN
New Features in PSP2 for SANsymphony -V10 Software-defined Storage Platform and DataCore Virtual SAN Updated: May 19, 2015 Contents Introduction... 1 Cloud Integration... 1 OpenStack Support... 1 Expanded
Let s Talk About Storage & Recovery Methods for Non-Volatile Memory Database Systems
Let s Talk About Storage & Recovery Methods for Non-Volatile Database Systems Joy Arulraj [email protected] Andrew Pavlo [email protected] Subramanya R. Dulloor [email protected] Carnegie
Cumulus: filesystem backup to the Cloud
Michael Vrable, Stefan Savage, a n d G e o f f r e y M. V o e l k e r Cumulus: filesystem backup to the Cloud Michael Vrable is pursuing a Ph.D. in computer science at the University of California, San
Moving Virtual Storage to the Cloud
Moving Virtual Storage to the Cloud White Paper Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage www.parallels.com Table of Contents Overview... 3 Understanding the Storage
MinCopysets: Derandomizing Replication In Cloud Storage
MinCopysets: Derandomizing Replication In Cloud Storage Asaf Cidon, Ryan Stutsman, Stephen Rumble, Sachin Katti, John Ousterhout and Mendel Rosenblum Stanford University [email protected], {stutsman,rumble,skatti,ouster,mendel}@cs.stanford.edu
