A COOL AND PRACTICAL ALTERNATIVE TO TRADITIONAL HASH TABLES
|
|
|
- Amanda Wood
- 10 years ago
- Views:
Transcription
1 A COOL AND PRACTICAL ALTERNATIVE TO TRADITIONAL HASH TABLES ULFAR ERLINGSSON, MARK MANASSE, FRANK MCSHERRY MICROSOFT RESEARCH SILICON VALLEY MOUNTAIN VIEW, CALIFORNIA, USA ABSTRACT Recent advances in the theoretical literature have proposed interesting modifications to traditional hash tables. The authors of these papers propose hash tables which a) have a guaranteed constant time for a lookup b) have amortized constant time for an insertion c) require table size only slightly larger than the space for the elements Previous hash table technologies have offered at most two of these three. Moreover, these hash tables do no dynamic memory allocation (except when the tables have to be resized), are easy to code, admit efficient concurrent implementations, and can be tuned (even dynamically). This is enabled by one simple idea: allow elements in the hash table to have multiple potential locations by using multiple hash functions, and allow elements to be relocated among potential locations to make room for new elements. Such relocations can cascade, inducing a search-tree to find an empty slot, and a sequence of relocations to allow the insertion of a new key. INTRODUCTION Hash tables are a useful abstraction in many applications, allowing for highly efficient expected time to maintain an association between keys and data. Traditional hash tables provide for reasonably fast lookup, reasonably fast insertion, and reasonably efficient use of memory. Some known variants give up some of these to improve on others perfect hashing, for example, uses more memory to allow expected rapid table construction, and constant time lookup, while open-address hashing allows some lookups and insertions to be slow, while using little memory. Recent developments [1,2,3] have proposed a new class of hash tables which provide for highly space-efficient tables, offering constant-time lookup and amortized constant-time insertion. In this paper, we present a simple generalization of these schemes, and note that some of these are well-suited to modern memory hierarchies, and to distributed settings. 1
2 THE NEW ALGORITHM, BASIC VERSION In the simplest version of this scheme, a hash table is stored in an array of n elements. Each element e has a set of locations L e at which it might be found in the table. If L e consists of a single location, we have standard hashing. This version of standard hashing will fail to insert colliding elements; this can be expected to start happening when the number of elements inserted exceeds O( n). In other traditional schemes, we would insert an element at the first unoccupied location in L e ; this describes many variations on chained hashing, where the list of locations includes all possible overflow locations in the order in which we would probe them. In the new schemes, we consider L e s which contain at least two unrelated locations. If any of these locations are unoccupied, we use the first unoccupied location to hold element e. If none are vacant, we create a space, by finding an element g currently located in some location in L e which can be relocated to some different location in L g ; this may entail relocating a chain of other elements. More concretely, we find a list of elements g 1, g 2, g 3,, g k, such that: 1) There is a vacant location v L gk. 2) The current location of L gi is also in L gi+1. 3) L e contains the current location of L g1. Having found such a list, we attempt to move g k to v, each g i to the initial location of g i+1, and then (if these moves succeeded, which they will in the absence of concurrent updates) insert e at the initial location of g 1. If we cannot find a set of relocations in a reasonable amount of time, insertion fails. In the presence of concurrency, we don t need to undo any moves which succeeded, we merely try again to find a new list of g s (probably of different length) and repeat the attempt to make space. The techniques for finding a list of g s range from performing a breadth-first search of the elements currently occupying the positions in L e to performing a random walk, picking the element currently occupying a random position in L e. In the event that all L e are of size two, these two searches are the same; this is the algorithm proposed in [2]. If each L e has at least two unrelated locations, and at least three locations, the tree will have a branching factor of at least two. In any case, lookup is straightforward: check each location in L e and see if e is there. If all the L e are of some small maximum size, this is manifestly a constant-time operation. PREVIOUS WORK If each L e has exactly two independently chosen random locations, the construction of the previous yields Pagh and Rodler s Cuckoo Hashing technique [2]. The algorithm works in this setting, but without good space efficiency; once the table is about half full (that is, roughly n/2 elements have been inserted), insertions are likely to start failing. This is due to a threshold property of random graphs: each of the sets can be viewed as an edge in a graph. The theory of random graphs is wellunderstood: most properties exhibit a threshold, where random graphs with sufficiently low edge probability don t exhibit the property, but once the edge probability exceeds the threshold value, the graph is very likely to exhibit the property. In this case, the property is the existence of cycles, for which the threshold occurs at n/2 edges. Once cycles exist, they exist in profusion, and the largest cycles are quite large. When an edge is introduced linking two nodes already in the same cycle, an irresolvable conflict is produced; again, this happens when the table is 50% full. Other papers explore expanded sets of choices for location sets, which we describe below. 2
3 Fotokis et al [1] take L e to be of size 4, all independent, except that each is in a different quarter of the table. By doing this, they demonstrate a load factor of approximately 97% before the first irresolvable conflict is produced. They provide a theoretical justification for this. It should be recognized that this is quite good: with four selections for each element, each drawn from a table of size n/4, we expect that the probability that a cell has no elements which map to it is ~(1-4/n) n, which is approximately e -4, that is, roughly.02. We know that the load can t possibly exceed n without conflict, and so, even without considering the graph theory, we can t expect to exceed a lood of 98%. The paper further considers extensions to larger sets of random choices, to drive the load factor still higher. In contrast, Panigrahy [3] makes one additional simplifying assumption: he also chooses 4 locations, but assumes that the locations are pairwise related. To do so, we divide the array into bins of size 2, or bins with 2 cells, with locations 2x and 2x+1 forming bin x. By so doing, we reduce the number of hash computations needed to 2 instead of 4. Lookups also take advantage of locality; memory systems are more efficient when looking at consecutive memory locations, due to caches, prefetching, and reduced pressure on the translation look-aside buffers (TLBs). In so doing, he proves an attainable load factor of at least 87%; that is, below that 87% load, we re almost certain to be able to insert all of our elements into the table, and above that we re very likely to encounter an irresolvable collision. This paper then considers increasing the size of bins, demonstrating theoretically an increase in attainable load factor. We see that four independent locations are superior to two pairs of independent locations, but impose a greater burden, increasing the computation of pseudo-random values, and adding pressure on the memory system. WHAT HAVE WE DONE? The two papers mentioned above extend cuckoo hashing in different ways: one extends the number of hash functions used (and, to simplify the analysis but with impaired capacity, splits the table into disjoint ranges for different hash functions); the other suggests using bins of size larger than 1, and two hash functions. We can unify these by considering cuckoo hashing using a two-parameter family: a bin size, and a set of hash functions. Standard cuckoo hashing results from the parameter choice (1, 2), that is a bin of size one, and two hash functions. Fotokis-style hashing is, in this parametrization, (1, 4) (or more generally, (1, d)). Panigrahy-style hashing is (2, 2) (or more generally (k, 2)). We have experimentally tried a variety of choices; (2, 3), for instance, achieves the same load factor as (1, 4) but with one fewer hash function evaluation, and Number of hash fu ncti ons 4 hash func 3 hash func 2 hash func 1 hash func 97% 91% 49% 0.06% 99% 97% 86% 0.6% 99.9% 98% 93% 12% (assuming that a length 2 bin fits in a cache line, as it should for reasonable element sizes) fewer cache misses. We have not considered these choices analytically, except to observe that the existing proofs serve to establish lower bounds. We also note that increasing load sets a lower bound on the 3% 99.9% 96% 1 cell 2 cells 4 cells 8 cells Number of cells per hash bucket Figure 1: Achievable load, 99% of the time, with 64K cells and fixed amount of work per insert 3
4 expected number of locations that must be inspected to complete an insertion. We favor setting the bin size to be as large as practical without increasing pressure on the memory hierarchy stressed by a given table size. For tables that fit in memory, but not in L1 cache, this is likely to be the size of a cache line, and aligned with the cache line boundaries. This may not always be true: sufficiently large tables in user space will mean that most access to memory will cause a TLB fault; it may be profitable to increase the size of a bin to amortize the cost of that as well. For very large tables, which we expect to be disk-resident, it may make sense to increase the bin size to fill the expected unit of reading and writing to disk. Nominally, this is a sector, but current disks and controllers typically make considerably larger operations efficient. asdf asdf Figure 2: Prob[couldn t achieve load] for (1,3) Figure 3: Amount of work per insert for (1,3) The three figures above show some results from our experiments. Figure 1 reports on tables with 64K cells, while Figures 2 and 3 report on tables with 4K cells. As the latter two figures demonstrate, for any given L e, the amount of work per inserted key remains reasonable until high load, and there is a sharp cutoff load at which point it becomes infeasible to insert additional keys. The total work required to fill this type of hash table up to its supported load, is amortized constant per inserted key. Experimentally, we find (8, 2) to be almost as space-efficient as (1,4) and (2, 3); if elements can be represented in 8 bytes, this requires a line size of 64 bytes, which matches a Pentium 4 L1 cache line. We have also considered the impact of allowing correlated-but-misaligned consolidation of elements. In this variant, each hash function would map into the full range, and the probability of unused locations is reduced, but we haven t run the experiments. That is, bins would overlap, with the advantage that fewer cells would be unreachable due to having no hashes evaluate to a particular bin, but the advantages of cache line reuse would be diminished. In a distributed setting, or for disk-based hash tables, it would also be interesting to consider much larger bins, related, say, to the size of disk pages or network transmissions. Interesting variants for the distributed setting might also want to consider hash functions which make it more likely that elements will be relocated to other bins residing on the same host, to reduce network traffic during relocation. WHY IS THIS USEFUL? Many applications use hash tables. In the Web-focused research we do, we run experiments on extremely large collections of web pages, where we use hash tables to store information about each server, website, or page. In some cases we track the collection of phrases on pages, looking for collisions between documents. In such cases, our data sets place billions, or even trillions, of entries into our hash tables. In these cases, our performance is highly limited by the memory hierarchy: tables that fit in the memory of a computer are much faster than those which have to access disk 4
5 frequently. Compact representations of the hash tables are a boon. Databases use hash tables to implement hash joins, at the very least, so this may come in handy there. Lots of programs want to keep a bounded size cache of recent actions. Using these hash tables, we can make that cache associative, and keep as large a set as possible in a fixed amount of memory. WHY ELSE IS THIS COOL? These hash tables are highly robust. For instance, a hash table constructed using a small number of hash functions is still valid if viewed with more hash functions; this means that we can dynamically increase the number of hash functions to stretch the capacity of the table, if needed. Also, the tables can be valid whether keys are present in one or more of their possible locations, meaning that the relocation operation doesn t need to delete a key from an old location. Furthermore, as we move items, the table is always in a consistent state, as long as moving an item from one location to another is atomic (or even copying an item). The atomic actions are small: by finding a plan for moving elements, and then executing it in reverse the tables are always consistent, and, as described earlier, if we encounter a conflict in the middle of executing a plan, we just leave the elements where they are, and pick a new plan. We have started experimenting with optimizing lookups by trying to put elements where the first hash function points, for those elements where lookups are frequent. We probabilistically decide to try to improve the placement of elements, so that we don t need to track access frequency, but can just say that one-quarter of the time, we try to move an element to a more-preferred location. The hash tables are easily expanded, when the load approaches the maximum: add more table space, and expand the range of the hash functions to map into the old and new space, either by adding new hash functions mapping into the new space (or mapping to both old and new space), or by reinterpreting bins to include old and new locations. We can easily change hash functions, without fully rehashing the table all at once: expand the sets L e by adding new hash functions, but mark the old locations in L e as lookup-only, so that, over time, elements migrate to new hash locations. When all elements have migrated (we can keep a counter of the unmigrated elements, and trigger a forced migration when that counter becomes small enough) we can retire the lookup-only hash functions. We can continue to break the table space into bins, but use hash functions which select a random starting element of the bin, to make operations even faster most of the time. Having done this, it s now easy to grow the size of bins, should that seem to be a good idea: we can merge adjacent pairs of bins effortlessly. This can be used to make the initial insertion of elements faster, but to allow higher utilization as the table becomes full. CONCLUSION Few algorithms have proven themselves as broadly useful as hash tables. With over a half-century of history, it is remarkable that new avenues remain, offering practical improvements in performance and efficiency. 5
6 REFERENCES 1. D. Fotakis et al., Space Efficient Hash Tables With Worst Case Constant Access Time, in 20 th Annual Symposium on Theoretical Aspects of Computer Science, pp , Berlin, Germany, R. Pagh and F. F. Rodler, Cuckoo Hashing, in 9 th Annual European Symposium on Algorithms, v.2161 of Lecture Notes in Computer Science, pp , Springer-Verlag, R. Panigrahy, Efficient Hashing with Lookups in Two Memory Accesses, in 16 th Annual ACM-SIAM Symposium on Discrete Algorithms, pp , Vancouver, British Columbia, Canada,
Universal hashing. In other words, the probability of a collision for two different keys x and y given a hash function randomly chosen from H is 1/m.
Universal hashing No matter how we choose our hash function, it is always possible to devise a set of keys that will hash to the same slot, making the hash scheme perform poorly. To circumvent this, we
Storage Management for Files of Dynamic Records
Storage Management for Files of Dynamic Records Justin Zobel Department of Computer Science, RMIT, GPO Box 2476V, Melbourne 3001, Australia. [email protected] Alistair Moffat Department of Computer Science
Physical Data Organization
Physical Data Organization Database design using logical model of the database - appropriate level for users to focus on - user independence from implementation details Performance - other major factor
A Survey on Efficient Hashing Techniques in Software Configuration Management
A Survey on Efficient Hashing Techniques in Software Configuration Management Bernhard Grill Vienna University of Technology Vienna, Austria Email: [email protected] Abstract This paper presents
Lecture 17: Virtual Memory II. Goals of virtual memory
Lecture 17: Virtual Memory II Last Lecture: Introduction to virtual memory Today Review and continue virtual memory discussion Lecture 17 1 Goals of virtual memory Make it appear as if each process has:
CS 2112 Spring 2014. 0 Instructions. Assignment 3 Data Structures and Web Filtering. 0.1 Grading. 0.2 Partners. 0.3 Restrictions
CS 2112 Spring 2014 Assignment 3 Data Structures and Web Filtering Due: March 4, 2014 11:59 PM Implementing spam blacklists and web filters requires matching candidate domain names and URLs very rapidly
Fundamental Algorithms
Fundamental Algorithms Chapter 7: Hash Tables Michael Bader Winter 2014/15 Chapter 7: Hash Tables, Winter 2014/15 1 Generalised Search Problem Definition (Search Problem) Input: a sequence or set A of
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
Cuckoo Filter: Practically Better Than Bloom
Cuckoo Filter: Practically Better Than Bloom Bin Fan, David G. Andersen, Michael Kaminsky, Michael D. Mitzenmacher Carnegie Mellon University, Intel Labs, Harvard University {binfan,dga}@cs.cmu.edu, [email protected],
Multi-dimensional index structures Part I: motivation
Multi-dimensional index structures Part I: motivation 144 Motivation: Data Warehouse A definition A data warehouse is a repository of integrated enterprise data. A data warehouse is used specifically for
Practical Survey on Hash Tables. Aurelian Țuțuianu
Practical Survey on Hash Tables Aurelian Țuțuianu In memoriam Mihai Pătraşcu (17 July 1982 5 June 2012) I have no intention to ever teach computer science. I want to teach the love for computer science,
Chapter 13. Disk Storage, Basic File Structures, and Hashing
Chapter 13 Disk Storage, Basic File Structures, and Hashing Chapter Outline Disk Storage Devices Files of Records Operations on Files Unordered Files Ordered Files Hashed Files Dynamic and Extendible Hashing
Record Storage and Primary File Organization
Record Storage and Primary File Organization 1 C H A P T E R 4 Contents Introduction Secondary Storage Devices Buffering of Blocks Placing File Records on Disk Operations on Files Files of Unordered Records
Data Warehousing und Data Mining
Data Warehousing und Data Mining Multidimensionale Indexstrukturen Ulf Leser Wissensmanagement in der Bioinformatik Content of this Lecture Multidimensional Indexing Grid-Files Kd-trees Ulf Leser: Data
1. Comments on reviews a. Need to avoid just summarizing web page asks you for:
1. Comments on reviews a. Need to avoid just summarizing web page asks you for: i. A one or two sentence summary of the paper ii. A description of the problem they were trying to solve iii. A summary of
Operating Systems CSE 410, Spring 2004. File Management. Stephen Wagner Michigan State University
Operating Systems CSE 410, Spring 2004 File Management Stephen Wagner Michigan State University File Management File management system has traditionally been considered part of the operating system. Applications
Copyright 2007 Ramez Elmasri and Shamkant B. Navathe. Slide 13-1
Slide 13-1 Chapter 13 Disk Storage, Basic File Structures, and Hashing Chapter Outline Disk Storage Devices Files of Records Operations on Files Unordered Files Ordered Files Hashed Files Dynamic and Extendible
FPGA-based Multithreading for In-Memory Hash Joins
FPGA-based Multithreading for In-Memory Hash Joins Robert J. Halstead, Ildar Absalyamov, Walid A. Najjar, Vassilis J. Tsotras University of California, Riverside Outline Background What are FPGAs Multithreaded
Database Systems. Session 8 Main Theme. Physical Database Design, Query Execution Concepts and Database Programming Techniques
Database Systems Session 8 Main Theme Physical Database Design, Query Execution Concepts and Database Programming Techniques Dr. Jean-Claude Franchitti New York University Computer Science Department Courant
Chapter 13. Chapter Outline. Disk Storage, Basic File Structures, and Hashing
Chapter 13 Disk Storage, Basic File Structures, and Hashing Copyright 2007 Ramez Elmasri and Shamkant B. Navathe Chapter Outline Disk Storage Devices Files of Records Operations on Files Unordered Files
Chapter 13 Disk Storage, Basic File Structures, and Hashing.
Chapter 13 Disk Storage, Basic File Structures, and Hashing. Copyright 2004 Pearson Education, Inc. Chapter Outline Disk Storage Devices Files of Records Operations on Files Unordered Files Ordered Files
18-548/15-548 Associativity 9/16/98. 7 Associativity. 18-548/15-548 Memory System Architecture Philip Koopman September 16, 1998
7 Associativity 18-548/15-548 Memory System Architecture Philip Koopman September 16, 1998 Required Reading: Cragon pg. 166-174 Assignments By next class read about data management policies: Cragon 2.2.4-2.2.6,
SMALL INDEX LARGE INDEX (SILT)
Wayne State University ECE 7650: Scalable and Secure Internet Services and Architecture SMALL INDEX LARGE INDEX (SILT) A Memory Efficient High Performance Key Value Store QA REPORT Instructor: Dr. Song
Comp 5311 Database Management Systems. 16. Review 2 (Physical Level)
Comp 5311 Database Management Systems 16. Review 2 (Physical Level) 1 Main Topics Indexing Join Algorithms Query Processing and Optimization Transactions and Concurrency Control 2 Indexing Used for faster
Memory management basics (1) Requirements (1) Objectives. Operating Systems Part of E1.9 - Principles of Computers and Software Engineering
Memory management basics (1) Requirements (1) Operating Systems Part of E1.9 - Principles of Computers and Software Engineering Lecture 7: Memory Management I Memory management intends to satisfy the following
EFFICIENT EXTERNAL SORTING ON FLASH MEMORY EMBEDDED DEVICES
ABSTRACT EFFICIENT EXTERNAL SORTING ON FLASH MEMORY EMBEDDED DEVICES Tyler Cossentine and Ramon Lawrence Department of Computer Science, University of British Columbia Okanagan Kelowna, BC, Canada [email protected]
Data Structures Fibonacci Heaps, Amortized Analysis
Chapter 4 Data Structures Fibonacci Heaps, Amortized Analysis Algorithm Theory WS 2012/13 Fabian Kuhn Fibonacci Heaps Lacy merge variant of binomial heaps: Do not merge trees as long as possible Structure:
each college c i C has a capacity q i - the maximum number of students it will admit
n colleges in a set C, m applicants in a set A, where m is much larger than n. each college c i C has a capacity q i - the maximum number of students it will admit each college c i has a strict order i
Accelerate Cloud Computing with the Xilinx Zynq SoC
X C E L L E N C E I N N E W A P P L I C AT I O N S Accelerate Cloud Computing with the Xilinx Zynq SoC A novel reconfigurable hardware accelerator speeds the processing of applications based on the MapReduce
File-System Implementation
File-System Implementation 11 CHAPTER In this chapter we discuss various methods for storing information on secondary storage. The basic issues are device directory, free space management, and space allocation
The Deadlock Problem. Deadlocks. Deadlocks. Bridge Crossing Example
The Deadlock Problem Deadlocks A set of blocked processes each holding a resource and waiting to acquire a resource held by another process in the set. Example System has 2 tape drives. P 1 and P 2 each
CURTAIL THE EXPENDITURE OF BIG DATA PROCESSING USING MIXED INTEGER NON-LINEAR PROGRAMMING
Journal homepage: http://www.journalijar.com INTERNATIONAL JOURNAL OF ADVANCED RESEARCH RESEARCH ARTICLE CURTAIL THE EXPENDITURE OF BIG DATA PROCESSING USING MIXED INTEGER NON-LINEAR PROGRAMMING R.Kohila
DATABASE DESIGN - 1DL400
DATABASE DESIGN - 1DL400 Spring 2015 A course on modern database systems!! http://www.it.uu.se/research/group/udbl/kurser/dbii_vt15/ Kjell Orsborn! Uppsala Database Laboratory! Department of Information
A Comparison of Dictionary Implementations
A Comparison of Dictionary Implementations Mark P Neyer April 10, 2009 1 Introduction A common problem in computer science is the representation of a mapping between two sets. A mapping f : A B is a function
Speeding Up Cloud/Server Applications Using Flash Memory
Speeding Up Cloud/Server Applications Using Flash Memory Sudipta Sengupta Microsoft Research, Redmond, WA, USA Contains work that is joint with B. Debnath (Univ. of Minnesota) and J. Li (Microsoft Research,
Tables so far. set() get() delete() BST Average O(lg n) O(lg n) O(lg n) Worst O(n) O(n) O(n) RB Tree Average O(lg n) O(lg n) O(lg n)
Hash Tables Tables so far set() get() delete() BST Average O(lg n) O(lg n) O(lg n) Worst O(n) O(n) O(n) RB Tree Average O(lg n) O(lg n) O(lg n) Worst O(lg n) O(lg n) O(lg n) Table naïve array implementation
Scalable Prefix Matching for Internet Packet Forwarding
Scalable Prefix Matching for Internet Packet Forwarding Marcel Waldvogel Computer Engineering and Networks Laboratory Institut für Technische Informatik und Kommunikationsnetze Background Internet growth
Lecture 1: Data Storage & Index
Lecture 1: Data Storage & Index R&G Chapter 8-11 Concurrency control Query Execution and Optimization Relational Operators File & Access Methods Buffer Management Disk Space Management Recovery Manager
Online Bipartite Perfect Matching With Augmentations
Online Bipartite Perfect Matching With Augmentations Kamalika Chaudhuri, Constantinos Daskalakis, Robert D. Kleinberg, and Henry Lin Information Theory and Applications Center, U.C. San Diego Email: [email protected]
Reconfigurable Architecture Requirements for Co-Designed Virtual Machines
Reconfigurable Architecture Requirements for Co-Designed Virtual Machines Kenneth B. Kent University of New Brunswick Faculty of Computer Science Fredericton, New Brunswick, Canada [email protected] Micaela Serra
Original-page small file oriented EXT3 file storage system
Original-page small file oriented EXT3 file storage system Zhang Weizhe, Hui He, Zhang Qizhen School of Computer Science and Technology, Harbin Institute of Technology, Harbin E-mail: [email protected]
Fast and Compact Hash Tables for Integer Keys
Fast and Compact Hash Tables for Integer Keys Nikolas Askitis School of Computer Science and Information Technology, RMIT University, Melbourne 3001, Australia. Email: [email protected] Abstract
MAD2: A Scalable High-Throughput Exact Deduplication Approach for Network Backup Services
MAD2: A Scalable High-Throughput Exact Deduplication Approach for Network Backup Services Jiansheng Wei, Hong Jiang, Ke Zhou, Dan Feng School of Computer, Huazhong University of Science and Technology,
Asking Hard Graph Questions. Paul Burkhardt. February 3, 2014
Beyond Watson: Predictive Analytics and Big Data U.S. National Security Agency Research Directorate - R6 Technical Report February 3, 2014 300 years before Watson there was Euler! The first (Jeopardy!)
COSC 6374 Parallel Computation. Parallel I/O (I) I/O basics. Concept of a clusters
COSC 6374 Parallel Computation Parallel I/O (I) I/O basics Spring 2008 Concept of a clusters Processor 1 local disks Compute node message passing network administrative network Memory Processor 2 Network
COS 318: Operating Systems. Virtual Memory and Address Translation
COS 318: Operating Systems Virtual Memory and Address Translation Today s Topics Midterm Results Virtual Memory Virtualization Protection Address Translation Base and bound Segmentation Paging Translation
Improving the Database Logging Performance of the Snort Network Intrusion Detection Sensor
-0- Improving the Database Logging Performance of the Snort Network Intrusion Detection Sensor Lambert Schaelicke, Matthew R. Geiger, Curt J. Freeland Department of Computer Science and Engineering University
Previous Lectures. B-Trees. External storage. Two types of memory. B-trees. Main principles
B-Trees Algorithms and data structures for external memory as opposed to the main memory B-Trees Previous Lectures Height balanced binary search trees: AVL trees, red-black trees. Multiway search trees:
BM307 File Organization
BM307 File Organization Gazi University Computer Engineering Department 9/24/2014 1 Index Sequential File Organization Binary Search Interpolation Search Self-Organizing Sequential Search Direct File Organization
How To Encrypt Data With A Power Of N On A K Disk
Towards High Security and Fault Tolerant Dispersed Storage System with Optimized Information Dispersal Algorithm I Hrishikesh Lahkar, II Manjunath C R I,II Jain University, School of Engineering and Technology,
CS514: Intermediate Course in Computer Systems
: Intermediate Course in Computer Systems Lecture 7: Sept. 19, 2003 Load Balancing Options Sources Lots of graphics and product description courtesy F5 website (www.f5.com) I believe F5 is market leader
Union-Find Algorithms. network connectivity quick find quick union improvements applications
Union-Find Algorithms network connectivity quick find quick union improvements applications 1 Subtext of today s lecture (and this course) Steps to developing a usable algorithm. Define the problem. Find
Data storage Tree indexes
Data storage Tree indexes Rasmus Pagh February 7 lecture 1 Access paths For many database queries and updates, only a small fraction of the data needs to be accessed. Extreme examples are looking or updating
Operating Systems, 6 th ed. Test Bank Chapter 7
True / False Questions: Chapter 7 Memory Management 1. T / F In a multiprogramming system, main memory is divided into multiple sections: one for the operating system (resident monitor, kernel) and one
Binary Heap Algorithms
CS Data Structures and Algorithms Lecture Slides Wednesday, April 5, 2009 Glenn G. Chappell Department of Computer Science University of Alaska Fairbanks [email protected] 2005 2009 Glenn G. Chappell
Krishna Institute of Engineering & Technology, Ghaziabad Department of Computer Application MCA-213 : DATA STRUCTURES USING C
Tutorial#1 Q 1:- Explain the terms data, elementary item, entity, primary key, domain, attribute and information? Also give examples in support of your answer? Q 2:- What is a Data Type? Differentiate
CS 464/564 Introduction to Database Management System Instructor: Abdullah Mueen
CS 464/564 Introduction to Database Management System Instructor: Abdullah Mueen LECTURE 14: DATA STORAGE AND REPRESENTATION Data Storage Memory Hierarchy Disks Fields, Records, Blocks Variable-length
Job Reference Guide. SLAMD Distributed Load Generation Engine. Version 1.8.2
Job Reference Guide SLAMD Distributed Load Generation Engine Version 1.8.2 June 2004 Contents 1. Introduction...3 2. The Utility Jobs...4 3. The LDAP Search Jobs...11 4. The LDAP Authentication Jobs...22
Big Data Technology Map-Reduce Motivation: Indexing in Search Engines
Big Data Technology Map-Reduce Motivation: Indexing in Search Engines Edward Bortnikov & Ronny Lempel Yahoo Labs, Haifa Indexing in Search Engines Information Retrieval s two main stages: Indexing process
Distributed File Systems
Distributed File Systems Paul Krzyzanowski Rutgers University October 28, 2012 1 Introduction The classic network file systems we examined, NFS, CIFS, AFS, Coda, were designed as client-server applications.
Deployment of express checkout lines at supermarkets
Deployment of express checkout lines at supermarkets Maarten Schimmel Research paper Business Analytics April, 213 Supervisor: René Bekker Faculty of Sciences VU University Amsterdam De Boelelaan 181 181
Virtual Memory Paging
COS 318: Operating Systems Virtual Memory Paging Kai Li Computer Science Department Princeton University (http://www.cs.princeton.edu/courses/cos318/) Today s Topics Paging mechanism Page replacement algorithms
Overlapping Data Transfer With Application Execution on Clusters
Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm [email protected] [email protected] Department of Computer Science Department of Electrical and Computer
CSE 326, Data Structures. Sample Final Exam. Problem Max Points Score 1 14 (2x7) 2 18 (3x6) 3 4 4 7 5 9 6 16 7 8 8 4 9 8 10 4 Total 92.
Name: Email ID: CSE 326, Data Structures Section: Sample Final Exam Instructions: The exam is closed book, closed notes. Unless otherwise stated, N denotes the number of elements in the data structure
Guideline for stresstest Page 1 of 6. Stress test
Guideline for stresstest Page 1 of 6 Stress test Objective: Show unacceptable problems with high parallel load. Crash, wrong processing, slow processing. Test Procedure: Run test cases with maximum number
Chapter 12 File Management
Operating Systems: Internals and Design Principles, 6/E William Stallings Chapter 12 File Management Dave Bremer Otago Polytechnic, N.Z. 2008, Prentice Hall Roadmap Overview File organisation and Access
Chapter 12 File Management. Roadmap
Operating Systems: Internals and Design Principles, 6/E William Stallings Chapter 12 File Management Dave Bremer Otago Polytechnic, N.Z. 2008, Prentice Hall Overview Roadmap File organisation and Access
Operating Systems. Virtual Memory
Operating Systems Virtual Memory Virtual Memory Topics. Memory Hierarchy. Why Virtual Memory. Virtual Memory Issues. Virtual Memory Solutions. Locality of Reference. Virtual Memory with Segmentation. Page
Benchmarking Hadoop & HBase on Violin
Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages
Using Synology SSD Technology to Enhance System Performance Synology Inc.
Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...
Load Balancing. Load Balancing 1 / 24
Load Balancing Backtracking, branch & bound and alpha-beta pruning: how to assign work to idle processes without much communication? Additionally for alpha-beta pruning: implementing the young-brothers-wait
Magento & Zend Benchmarks Version 1.2, 1.3 (with & without Flat Catalogs)
Magento & Zend Benchmarks Version 1.2, 1.3 (with & without Flat Catalogs) 1. Foreword Magento is a PHP/Zend application which intensively uses the CPU. Since version 1.1.6, each new version includes some
find model parameters, to validate models, and to develop inputs for models. c 1994 Raj Jain 7.1
Monitors Monitor: A tool used to observe the activities on a system. Usage: A system programmer may use a monitor to improve software performance. Find frequently used segments of the software. A systems
The Real Challenges of Configuration Management
The Real Challenges of Configuration Management McCabe & Associates Table of Contents The Real Challenges of CM 3 Introduction 3 Parallel Development 3 Maintaining Multiple Releases 3 Rapid Development
Binary search tree with SIMD bandwidth optimization using SSE
Binary search tree with SIMD bandwidth optimization using SSE Bowen Zhang, Xinwei Li 1.ABSTRACT In-memory tree structured index search is a fundamental database operation. Modern processors provide tremendous
Whitepaper: performance of SqlBulkCopy
We SOLVE COMPLEX PROBLEMS of DATA MODELING and DEVELOP TOOLS and solutions to let business perform best through data analysis Whitepaper: performance of SqlBulkCopy This whitepaper provides an analysis
Chapter Objectives. Chapter 9. Sequential Search. Search Algorithms. Search Algorithms. Binary Search
Chapter Objectives Chapter 9 Search Algorithms Data Structures Using C++ 1 Learn the various search algorithms Explore how to implement the sequential and binary search algorithms Discover how the sequential
INTRODUCTION The collection of data that makes up a computerized database must be stored physically on some computer storage medium.
Chapter 4: Record Storage and Primary File Organization 1 Record Storage and Primary File Organization INTRODUCTION The collection of data that makes up a computerized database must be stored physically
Overview of Storage and Indexing
Overview of Storage and Indexing Chapter 8 How index-learning turns no student pale Yet holds the eel of science by the tail. -- Alexander Pope (1688-1744) Database Management Systems 3ed, R. Ramakrishnan
Static Analysis of Dynamic Properties - Automatic Program Verification to Prove the Absence of Dynamic Runtime Errors
Static Analysis of Dynamic Properties - Automatic Program Verification to Prove the Absence of Dynamic Runtime Errors Klaus Wissing PolySpace Technologies GmbH Argelsrieder Feld 22 82234 Wessling-Oberpfaffenhofen
A Survey on Deduplication Strategies and Storage Systems
A Survey on Deduplication Strategies and Storage Systems Guljar Shaikh ((Information Technology,B.V.C.O.E.P/ B.V.C.O.E.P, INDIA) Abstract : Now a day there is raising demands for systems which provide
6. Storage and File Structures
ECS-165A WQ 11 110 6. Storage and File Structures Goals Understand the basic concepts underlying different storage media, buffer management, files structures, and organization of records in files. Contents
Lecture 2 February 12, 2003
6.897: Advanced Data Structures Spring 003 Prof. Erik Demaine Lecture February, 003 Scribe: Jeff Lindy Overview In the last lecture we considered the successor problem for a bounded universe of size u.
The Advantages and Disadvantages of Network Computing Nodes
Big Data & Scripting storage networks and distributed file systems 1, 2, in the remainder we use networks of computing nodes to enable computations on even larger datasets for a computation, each node
Offline sorting buffers on Line
Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: [email protected] 2 IBM India Research Lab, New Delhi. email: [email protected]
CSE 326: Data Structures B-Trees and B+ Trees
Announcements (4//08) CSE 26: Data Structures B-Trees and B+ Trees Brian Curless Spring 2008 Midterm on Friday Special office hour: 4:-5: Thursday in Jaech Gallery (6 th floor of CSE building) This is
Platter. Track. Index Mark. Disk Storage. PHY 406F - Microprocessor Interfacing Techniques
Platter PHY 406F - icroprocessor Interfacing Techniques Disk Storage The major "permanent" storage medium for computers is, at present, generally magnetic media in the form of either magnetic tape or disks.
InfiniteGraph: The Distributed Graph Database
A Performance and Distributed Performance Benchmark of InfiniteGraph and a Leading Open Source Graph Database Using Synthetic Data Objectivity, Inc. 640 West California Ave. Suite 240 Sunnyvale, CA 94086
File System Management
Lecture 7: Storage Management File System Management Contents Non volatile memory Tape, HDD, SSD Files & File System Interface Directories & their Organization File System Implementation Disk Space Allocation
Low-rate TCP-targeted Denial of Service Attack Defense
Low-rate TCP-targeted Denial of Service Attack Defense Johnny Tsao Petros Efstathopoulos University of California, Los Angeles, Computer Science Department Los Angeles, CA E-mail: {johnny5t, pefstath}@cs.ucla.edu
Chapter 1 File Organization 1.0 OBJECTIVES 1.1 INTRODUCTION 1.2 STORAGE DEVICES CHARACTERISTICS
Chapter 1 File Organization 1.0 Objectives 1.1 Introduction 1.2 Storage Devices Characteristics 1.3 File Organization 1.3.1 Sequential Files 1.3.2 Indexing and Methods of Indexing 1.3.3 Hash Files 1.4
APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM
152 APPENDIX 1 USER LEVEL IMPLEMENTATION OF PPATPAN IN LINUX SYSTEM A1.1 INTRODUCTION PPATPAN is implemented in a test bed with five Linux system arranged in a multihop topology. The system is implemented
Victor Shoup Avi Rubin. fshoup,[email protected]. Abstract
Session Key Distribution Using Smart Cards Victor Shoup Avi Rubin Bellcore, 445 South St., Morristown, NJ 07960 fshoup,[email protected] Abstract In this paper, we investigate a method by which smart
Using Synology SSD Technology to Enhance System Performance Synology Inc.
Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_WP_ 20121112 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges... 3 SSD
Abstract Data Type. EECS 281: Data Structures and Algorithms. The Foundation: Data Structures and Abstract Data Types
EECS 281: Data Structures and Algorithms The Foundation: Data Structures and Abstract Data Types Computer science is the science of abstraction. Abstract Data Type Abstraction of a data structure on that
