Memory System Characterization of Big Data Workloads

Size: px
Start display at page:

Download "Memory System Characterization of Big Data Workloads"

Transcription

1 2013 IEEE International Conference on Big Data Memory System Characterization of Big Data Workloads Martin Dimitrov*, Karthik Kumar*, Patrick Lu**, Vish Viswanathan*, Thomas Willhalm* *Software and Services Group, **Datacenter and Connected Systems Group, Intel Corporation Abstract Two recent trends that have emerged include (1) Rapid growth in big data technologies with new types of computing models to handle unstructured data, such as mapreduce and nosql (2) A growing focus on the memory subsystem for performance and power optimizations, particularly with emerging memory technologies offering different characteristics from conventional DRAM (bandwidths, read/write asymmetries). This paper examines how these trends may intersect by characterizing the memory access patterns of various Hadoop and nosql big data workloads. Using memory DIMM traces collected using special hardware, we analyze the spatial and temporal reference patterns to bring out several insights related to memory and platform usages, such as memory footprints, read-write ratios, bandwidths, latencies, etc. We develop an analysis methodology to understand how conventional optimizations such as caching, prediction, and prefetching may apply to these workloads, and discuss the implications on software and system design. Keywords-big data, memory characterization I. INTRODUCTION The massive information explosion over the past decade has resulted in zetabytes of data [1] being created each year, with most of this data being in the form of files, videos, logs, documents, images, etc (unstructured formats). Data continues to grow at an exponential pace, for example, in 2010, the world created half as much data as it had in all previous years combined [2]. With such data growth, it becomes challenging for conventional computing models to handle such large volumes. Big Data analytics [3] [4] [5] [6] has emerged as the solution to parse, analyze and extract meaningful information from these large volumes of unstructured data. Extracting this information provides opportunities and insights in a variety of fronts- from making more intelligent business decisions, understanding trends in usages and markets, detecting frauds and anomalies, etcwith many of these being possible real-time. As a result of these advantages, big data processing is becoming increasingly popular. Two primary big data computing models have emerged: (1) Hadoop-based computing, and (2) NoSQLbased computing, and the two are among the fastest growing software segments in modern computer systems. On the other hand, computing systems themselves have seen a shift in optimizations; unlike a decade earlier when most optimizations were primarily on the processor, there has been more focus on the overall platform, and particularly the memory subsystem for performance and power improvements. Recent studies have shown memory to be a firstorder consideration for performance, and sometimes even the dominant consumer of power in a computer system. This emphasis becomes even more important with recent trends in emerging memory technologies [7] [8], that are expected to offer different characteristics from conventional DRAMsuch as higher latencies, differing capacities, persistence, etc. In order to have software run efficiently using such technologies, it becomes critical to characterize and understand the memory usages.while various studies have been performed on memory characterization of workloads [9] [10] [11] [12] over the past decade, unfortunately, most of them focus on the SPEC benchmark suites and traditional computing models. Very few studies have examined memory behaviors of big data workloads- and these are mostly specific to an optimization, such as TLB improvements [13] [14]. This paper addresses this gap by providing a detailed characterization of the spatial and temporal memory references of various big data workloads. We analyze various building blocks of big data operations such as sort, wordcount, aggregations, and joins on Hadoop, and building indexes to process data on a NoSQL data store. We characterize the memory behavior by monitoring various metrics such as memory latencies, first, second and last level processor cache miss rates, code and data TLB miss rates, peak memory bandwidths, etc. We examine the impact of Hadoop compression both on performance and the spatial patterns. Using specially designed hardware, we are able to observe and trace the memory reference patterns of all the workloads at the DIMM level with precise timing information. These traces provide us the unique ability to obtain insights based on the spatial and temporal references, such as memory footprints, and the spatial histograms of the memory references over the footprints. This paper also examines the potential for big data workloads to be tolerant to the higher latencies expected in emerging memory technologies. The classic mechanisms for doing this are caching in a faster memory tier, and predicting future memory references to prefetch sections of memory. We examine the cacheability of big data workloads by running the memory traces through a cache simulator with different cache sizes. An interesting insight is that we observe is that many of the workloads operate on only a small subset of their spatial footprint at a time. As a result, we find that a cache that is less than 0.1% the size of the footprint can provide a hit rate as high as 40% of all the memory references. For prefetchability, we observe that using existing prefetcher schemes to predict the precise /13/$ IEEE 15

2 next memory reference is a hard problem at the memory DIMM level due to mixing of different streams from different processor cores, and this mixed stream getting further interleaved across different memory ranks for performance. Clearly, more sophisticated algorithms and techniques are required if prefetching is to be transparent to the application at the lower levels of the memory hierarchy. In order to examine the potential for the same, we use signal processing techniques: entropy and trend analysis (correlation with known signals) to bring out insights related to the memory patterns. We believe this is the first paper to examine the design space for memory architectures running big data workloads, by analyzing spatial patterns using DIMM traces and providing a detailed memory characterization. The study brings out a wealth of insights for system and software design. The experiments are performed on a 4 node cluster, using 2 socket servers with Intel Xeon E5 processors, with each node configured with 128GB of DDR3 memory, and 2TB of SSD storage. (We intentionally selected fast storage and large memory capacities: with the price for flash and non-volatile media continuing to drop, we chose this configuration to understand forward looking usages). The remainder of this paper is organized as follows: Section II describes the related work. Section III describes the various workloads used in this paper. Section IV describes the experimental methodology used, and Section V presents the results and observations, with Section VI concluding the paper and discussing future work. II. RELATED WORK A. Memory system characterization There have been several papers discussing memory system characterization of enterprise workloads, over the past decade. Barroso et al. [9] characterize the memory references of various commercial workloads. Domain specific characterization studies include memory characterization of parallel data mining workloads [15], of the ECperf benchmark [16], of memcached [17], and of the SPEC CPU2000 and CPU2006 benchmark suites [10] [11] [12]. Particularly noteworthy is the work of Shao et al. [12], where statistical measures are used for memory characterization. The common focus of all these works is using instrumentation techniques and platform monitoring to understand how specific workloads use memory. With emerging memory technologies [8] [18] [19] [20] having different properties from conventional DRAM that has been used for the past decade, this type of characterization focus becomes particularly important. B. Big Data workloads Various researchers have proposed benchmarks and workloads representative of big data usages; the common focus of all these benchmarks is they deal with processing unstructured data, typically using Hadoop or NoSQL. The Hibench suite developed by Intel [21] consists of several Hadoop workloads such as sort, wordcount, hive aggregation, etc. that are proxies for big data usages in the real world. In this paper, we use several Hadoop workloads from the HiBench suite. Another class of data stores to handle unstructured data are NoSQL databases [22], which are specialized for query and search operations. They differ from conventional databases in that they typically do not offer transactional guarantees, and this is a trade-off made in exchange for very fast retrieval. Recent studies have also proposed characterizing and understanding these big data usage cases. These can be classified as follows Implications on system design and architecture: A study from IBM research [23] examines how big data workloads may be suited for the IBM POWER architecture. Chang et al. [24] examine the implications of big data analytics on system design. Modeling big data workloads: Yang et al. [25] propose using statistics-based techniques for modeling map reduce. Atikoglu et al. [26] model and analyze the behavior of a key-value store (memcached). Performance characterization of big data workloads: Ren et al. characterize the behavior of a production Hadoop cluster, using a specific case study [27]. Issa et al. [28] present power and performance characterization of Hadoop with memcached. Very few studies focus on understanding the memory characteristics of big data workloads. Noteworthy among these are that of Basu et al. [14], which focuses on pagetable and virtual memory related optimizations for big data workloads. Jia et al. [13] present a characterization of L1, L2, and LLC cache misses observed for a Hadoop workload cluster. Both of these studies focus on characterization at the virtual memory and cache hierarchy, as opposed to the DRAM level. C. Contributions The following are the unique contributions of this paper: We believe this is the first study analyzing the memory reference patterns of big data workloads. Using hardware memory traces at the DIMM level, we are able to analyze references to physical memory. We introduce various metrics to qualitatively and quantitatively characterize the memory reference patterns, and we discuss the implications of these metrics for future system design. III. WORKLOADS We use several Hadoop workloads, from the HiBench workload suite [21] and a NoSQL datastore that builds indexes from text documents. We use a performance-optimized Hadoop configuration in our experiments. Since Hadoop has a compression codec for both input and output data, all the 16

3 Hadoop workloads are examined with and without use of compression. The following is a brief description of the various workloads used: A. Sort Sort is a good proxy of a common type of big data operation, that requires transforming data from one representation to another. In this case, the workload sorts its text input data, which is generated using the Hadoop RandomTextWriter example. In our setup, we sort a total of 96GB dataset in HDFS, using a 4 node cluster, with 24GB dataset/node. B. WordCount Word count also represents a common big data operation: extracting a small amount of interesting data from a large dataset, or a needle in haystack search. In this case, the workload counts the number of occurrences of each word in the input data set. The data set is generated using the Hadoop RandomTextWriter example. In our setup, we perform wordcount on a 128GB dataset in HDFS, distributed between the 4 nodes as 32GB per node. C. Hive Join The Hive join workload approximates a complex analytic query, representative of typical OLAP workloads. It computes the average and the sum for each group by joining two different tables. The join task consists of two sub-tasks that perform a complex calculation on two data sets. In the first part of the task, each system must find the sourceip that generated the most revenue within a particular date range. Once these intermediate records are generated, the system must then calculate the average pagerank of all the pages visited during this interval. The data set generated approximates web-server logs with hyperlinks following the Zipfian distribution. In this case, we simulated nearly 130 million user visits, to nearly 18 million pages. D. Hive Aggregation Hive aggregation approximates a complex analytic query, representative of typical OLAP workloads by computing the inlink count for each document in the dataset, a task that is often used as a component of PageRank calculations. The first step is to read each document and search for all the URLs that appear in the contents. The second step is then, for each unique URL, to count the number of unique pages that reference that particular URL across the entire set of documents. It is this type of task that the map-reduce is believed to be commonly used for. E. NoSQL indexing The NoSQL workload uses a NoSQL data store to build indexes, from 240GB of text files, distributed across the 4 nodes. This type of computation is heavy in regular expression comparisons, and is a very common big data use case. IV. EXPERIMENTAL METHODOLOGY In this section, we discuss the experimental methodology used in the paper. The experimental methodology is focused on the following objectives: (1) providing insights about how big data applications use the memory subsystem (Section IV-A) (2) examining the latency tolerance of big data workloads (since emerging memory technologies have higher latencies than DRAM). The latency tolerance is examined by understanding the potential for classic techniques to hide latency: cacheability in a faster tier (Sections IV-B, IV-C, and IV-D) and prefetching into a faster tier (IV-D and IV-E). A. General characterization Performance counter monitoring allows us to analyze various characteristics of the workload and its memory references. Some of these metrics that are of interest include: Memory footprints: the memory footprint is defined as the span of memory that is touched at least once by the workload. It can also be viewed as the minimum amount of memory required to keep the workload in memory. CPI: cycles per instruction. This is a measure of the average number of hardware cycles required to execute one instruction. A lower value indicates that the instructions are running efficiently on the hardware, with fewer stalls, dependencies, waits, and bottlenecks. L1, L2, Last Level Cache (LLC) misses per instruction (MPI): the processor has 3 levels of cache hierarchy. The L1 and L2 caches are smaller (kb sizes), fast caches that are exclusive to each core. The LLC is the last level cache that is shared amongst all the cores in a socket. Since the LLC is the last level in the hierarchy, the penalty for an LLC miss is a reference to DRAM, and this requires 10s of nanoseconds of wait time. Hence the LLC MPI is often a good indicator of the memory intensiveness of a workload. Memory bandwidth: Memory bandwidth is the data rate measured for references to the memory subsystem. Intel Xeon E5 two socket server platforms can easily support bandwidths greater than 60,000 MB/s. Instruction and Data Translation Lookaside Buffers (TLB) MPI: The TLBs are a cache for the page table entries. A higher miss rate for the data TLBs indicates memory references are more widespread in distribution since a TLB miss occurs whenever a 4kB boundary is crossed, and a page entry that is not cached in the TLB is referenced. B. Cache line working set characterization The spatial nature of the memory references of a workload can be identified by a characterization of the cache line references. For example, the memory referenced by a workload may span 5GB. However, it may be the case, that most of the references were concentrated on a smaller 17

4 100MB region within the 5GB. In order to understand such behavior, we employ the following methodology: (1) we create a histogram of the cache lines and their memory references: against each cache line, we have the number of times it is referenced (2) we sort the cache lines by their number of references, with the most frequently referenced cache line occurring first (3) we select a footprint size (for example: 100MB), and percentage of references contained within this footprint size. In this example, lines (=100MB/64 bytes per cache line) are required to contain 100MB; we compute the total number of references against the first cache lines in the list from step (2); and divide by the total number of references in the trace. This gives us information about the spatial distribution of the cache lines within the hottest 100MB, against the overall memory footprint. Intuitively, if one had a fixed cache and had to pin the cache lines with no replacements or additions possible, then the cache lines to be selected for this would be the ones highlighted in this analysis. C. Cache simulation The previous section considers the spatial distribution of the memory references; however, it does not consider the temporal nature. For example, the workload could be repeatedly streaming through memory in very large scans, that are repeated. However, if the working set spanned by the scan is much larger than the size of the cache, this will result in very poor cacheability. Moreover, there could be certain cache eviction and replacement patterns that could result in poor cacheability that is not apparent from inspecting the spatial distribution. On the other hand, it could be also possible that the workload focuses on small regions of memory at a time, resulting in very good cacheability; again this may not be apparent from the spatial distribution. In order to account for such cases, we run the memory reference traces through a cache simulator with different cache sizes, and observe the corresponding hit rates. High hit rates indicate a tiered memory architecture, such as a first tier of DRAM could be used to cache a good portion of the memory references to a second larger tier, based on a non-volatile technology. D. Entropy The previous metrics provide information about the spatial distribution, and how the temporal pattern impacts cacheability. Another important consideration for memorylevel optimizations is predictability and compressibility. This is related to the information content based on the observation that a signal with a high amount of information content is harder to compress, and potentially difficult to predict. In order to quantify and compare this feature for the different traces, we use entropy of the memory references as a metric, as it has been used in [12] as a metric for understanding memory reference patterns of SPEC workloads. The entropy is measure of the information content in the trace, and therefore gives good indication for its cacheability and predictability. For a set of cache lines K, the entropy is defined as H = c K p(c) log(p(c)) where p(c) denotes the probability of the cache line c. The following illustrative example demonstrates how the entropy can be used to characterize memory references: Consider a total of 10 distinct cache lines: {a, b, c, d, e, f, g, h, i, j} that are referenced in a trace and the following three scenarios, each consisting of 100 references to these cache lines: (1) Each of the 10 cache lines are referenced 10 times. (2) Cache lines a, b, c, d, e are referenced 19 times each, and cache lines f, g, h, i, j are referenced 1 time each. (3) Cache line a is referenced 91 times, and cache lines b, c, d, e, f, g, h, i, j are referenced 1 time each. All three access patterns use all 10 cache lines and have 100 cache line references. Metrics like footprint and reference counts therefore become identical in all the 3 cases. However, in the last case a single cache line contains 91% of the references, but only 19% of the references in case (2) and 10% of the references in case (1). Similarly, a set of 3 cache lines contains 93% of the references in case (3), 57% of the references in case (2) and 30% of the references in case (1). Therefore from a footprint or working set point of view, (3) is preferable over (2), which again is preferable over (1). This is nicely reflected in the entropy, which is 1 for scenario (1), for scenario (2), and for scenario (3). The number of references to various cache lines in a trace gets converted to a probability distribution between the cache lines in a trace, and is therefore relative between the cache lines. In particular, the entropy is independent of the length of a trace. E. Correlation and trend analysis To further understand the predictability of memory references, we examine the traces for trends. For example, if we knew that a trace had a trend of increasing physical address ranges in its references, aggressively prefetching large sections in the upper physical address ranges to an upper level cache would result in a fair percentage of cache hits. In order to quantify this, we use correlation analysis with known signals. The computation can be mathematically expressed as follows: c(n) = f g(n) = m= f[m]g[m + n] (1) Here g is the trace, and f is a known signal. For our 18

5 analysis, we use a single sawtooth function: { s n if 0 n l f s,l (n) = 0 otherwise The correlation output would then examine the trace g, looking for the known signal f. With a slope of s = 64 and a length of l = 1000, the test function f is mimicking a ascending stride through memory of 1000 cache lines. Please note that the infinite sum in (1) collapses to f s,l g(n) = l f s,l [m]g[m + n] = m=0 (2) l s m g[m + n] m=0 Furthermore, it is worth noting that test function for an descending stride with a negative slope s simply results in a negative correlation. f s,l g = ( f s,l ) g = f s,l g Figure 1. Memory footprints of the workloads V. RESULTS A. Experimental Setup We perform our experiments on a 4 node cluster, with each node being an Intel Xeon E5 2.9GHz two socket server platform, with 128GB 1600MHz DDR3 memory, and each node having 2TB of SSD storage. One of the nodes is fitted with the special hardware required to collect memory traces of the workloads at the DIMM level. The memory tracer interposes between the DIMM and the motherboard and is completely unobtrusive, while at the same time capable of recording all signals arriving at the pins of the memory DIMM. We keep the cluster configuration identical, and ensured and verified during our experiments that Hadoop and the NoSQL workload distribute tasks equally among the nodes; hence observations made at one node can be generalized. In order to collect the memory traces, we use specially designed hardware that can record physical address referenced at the DIMM level, without any overhead. B. General characterization In this section, we describe the memory characteristics of all the workloads. Memory footprints: Figure 1 shows the memory footprints of the workloads, in GB. It is observed that most workloads have footprints of 10GB or greater, with the NoSQL workload and the uncompressed sort workload having the largest footprints. It can also be observed that compression reduces the memory footprints, and helps reduce the execution time, as seen in Figure 3. Also, the nature of the footprints is mostly read-intensive, with the NoSQL workload and wordcount map having read-write ratios of 2 or greater. Among the Hadoop workloads, Hive join and Hive aggregation were found to have more writes, when compared to the sort Figure 2. Cycles per Instruction to execute the workloads Figure 3. Execution times for the workloads Figure 4. First level data cache misses per 1000 instructions 19

6 Figure 5. Second level cache misses per 1000 instructions Figure 9. Instruction TLB misses per 1000 instructions Figure 6. Last level cache misses per 1000 instructions Figure 7. Peak memory bandwidths recorded for the workloads and wordcount workloads. In most cases, we observed that enabling Hadoop compression reduced the read-write ratio. CPI: The CPI of the workloads is shown in Figure 2. It is observed that most of the workloads have CPI closer to 1, with uncompressed sort having the largest CPI. This difference with compression for the sort workload is also apparent in the execution times, shown in Figure 3. From all the Figures 2, 3, 1, it can be observed that the sort and wordcount workloads benefit most from using compression, followed by the Hive aggregation workload. L1, L2, LLC MPI: Figures 4, 5, and 6 show the miss rates per thousand instructions of the three levels of the cache hierarchy respectively. The sort workload is observed to have the highest cache miss rates. Intuitively, this makes sense because it transforms a large volume of data from one representation to another. The benefits of compression is also apparent in the last level cache miss rates of all the workloads. Memory bandwidth: The peak memory bandwidths of the workloads are shown in Figure 7. It is observed that all the workloads have peak bandwidths of several 10s of GB/s, all within the platform capability of 70 GB/s. Wordcount is observed to be the most bandwidth intensive amongst the workloads. We note that while some workloads have higher bandwidth with compression enabled, the total data traffic to memory (product of execution time and bandwidth) is lowered in all cases, with compression enabled. Instruction and Data TLB MPI: It is interesting to note that although the sort workload has almost an order of magnitude larger footprint than the wordcount workload, wordcount has much higher Data TLB miss rates. This indicates that the memory references of the workload are not well contained within page granularities, and are more widespread. In terms of instruction TLBs, the NoSQL workload is observed to have the highest miss rates. Figure 8. Data TLB misses per 1000 instructions C. Cache line working set characterization Figures 11 shows the working set characterization described in the earlier section. It is observed that the hottest 20

7 Figure 10. Entropy of cache line references Figure 13. Correlation of traces with known signal; suffixes are as follows p: prefetch, no: no prefetch, c: compression, nc: no compression Figure 11. Figure 12. Percentage references contained in memory footprints Cache miss rates for different cache sizes 100MB of cache lines contain 20% of the workloads for all, except for the NoSQL workload and the map phase of the uncompressed word count workload. The NoSQL workload stands out from the other workloads in terms of this characterization for the various footprint sizes. A 1GB footprint is observed to contain 60% of the memory references of all but this workload. It is also interesting to note that even though the sort workload has footprints of more than 100GB, more than 60% of its memory references are contained in 1GB, i.e. less than 1% of its footprint. D. Cache simulation Figure 12 shows the cacheability while accounting for the temporal nature of the reference patterns. It is interesting to compare Figure 11 and Figure 12 - and observe that the percentage of cache hits for the references is higher in Figure 12. This indicates that all the big data workloads do not operate on the entire footprint at once; rather they operate on spatial subsets of the total footprint, which makes them the hit rates higher in a cache that allows for evictions and replacements. It is observed that a 100MB cache has hit rates of 40-50% for most of the workloads, and a 1GB cache has hit rates of 80% for most workloads; indicating that these workloads are extremely cache friendly. Observing the trends- it is interesting to observe that the NoSQL workload appears to have lowest hit rates and slopes for both the working set analysis and the cache simulations. E. Entropy Figure 10 shows the entropies of all the workloads cache line references. It is observed that most of the big data workloads have entropies ranging from 13 to 16, with the NoSQL and sort workloads having the highest entropiesindicating large information content, harder predictability, and poorer compressibility. A common feature between these workloads is that they operate on entire datasets: both the inputs and outputs are comparably of the same size, with large transforms being performed. A noteworthy comparison for Figure 10 would be with the entropies for the SPEC workloads in [12]; it is observed that most of the SPEC workloads have entropies in the range of 8-13 (lower than the big data workloads), with equake and mcf being the only workloads to have entropies close to 16. F. Correlation and trend analysis Figure 13 shows the correlation analysis described in Section IV-E, with a known signal that has an increasing slope of 64, for some of the workloads. We normalize the correlation analysis using the mean and standard deviation to ensure fair comparisons can be made between the different workloads. A higher magnitude for the correlation indicates the trend (known signal) is observed strongly in the trace, with a positive magnitude denoting successive physical addresses are likely to be increasing in magnitude (by 64 bytes) and negative magnitude indicating successive physical addresses are likely to decrease in magnitude. It is observed that the Hive aggregation workload overall has high correlation magnitudes, indicating it may be beneficial to predict and prefetch memory references in the higher address ranges (when compression is disabled) and lower address ranges 21

8 (when compression is enabled). In most cases (other than the NoSQL workload), enabling prefetchers during trace collection results in higher correlation, as would be expected due to the prefetcher hitting adjacent addresses. In the case of the NoSQL workload, on further examination of the trace, we observed there were several local phases of increasing and decreasing trends that changed along the duration of the trace. VI. CONCLUSION AND OUTLOOK We examine the design space for memory architectures running big data workloads, by analyzing spatial patterns using DIMM traces and providing a detailed memory characterization and highlight various observations and insights for system design. Our study shows that several big data workloads can potentially hide latencies by caching references in a faster tier of memory. Moreover, there are trends (increasing address ranges) observable in these workloads, indicating potential for aggressively prefetching large sections of the dataset onto a faster tier. For future work, we plan to expand the measurements to include more big data workloads as well as exploring further ways to characterize the workloads, with variations in dataset size, etc. REFERENCES [1] Wikibon, Big data statistics, in [2] SeekingAlpha, Opportunities to play the explosive growth of big data, in opportunities-to-play-the-explosive-growth-of-big-data. [3] D. Boyd and K. Crawford, Six provocations for big data, Oxford Internet Institute: A Decade in Internet Time: Symposium on the Dynamics of the Internet and Society, [4] B. Brown, M. Chui, and J. Manyika, Are you ready for the era of big data? McKinsey Quarterly, vol. 4, pp , [5] K. Bakshi, Considerations for big data: Architecture and approach, in IEEE Aerospace Conference, 2012, pp [6] J. Manyika, M. Chui, B. Brown, J. Bughin, R. Dobbs, C. Roxburgh, and A. H. Byers, Big data: The next frontier for innovation, competition and productivity, Technical report, McKinsey Global Institute, Tech. Rep., [7] Y. Xie, Modeling, architecture, and applications for emerging memory technologies, Design & Test of Computers, IEEE, vol. 28, no. 1, pp , [8] A. Makarov, V. Sverdlov, and S. Selberherr, Emerging memory technologies: Trends, challenges, and modeling methods, Microelectronics Reliability, vol. 52, no. 4, pp , [9] L. A. Barroso, K. Gharachorloo, and E. Bugnion, Memory system characterization of commercial workloads, ACM SIGARCH Computer Architecture News, vol. 26, no. 3, pp. 3 14, [10] A. Jaleel, Memory characterization of workloads using instrumentation-driven simulation a pin-based memory characterization of the spec cpu2000 and spec cpu2006 benchmark suites, Intel Corporation, VSSAD, [11] F. Zeng, L. Qiao, M. Liu, and Z. Tang, Memory performance characterization of spec cpu2006 benchmarks using tsim, Physics Procedia, vol. 33, no. 0, pp , [12] Y. S. Shao and D. Brooks, Isa-independent workload characterization and its implications for specialized architectures, in IEEE International Symposium on Performance Analysis of Systems and Software, 2013, pp [13] Z. Jia, L. Wang, J. Zhan, L. Zhang, and C. Luo, Characterizing data analysis workloads in data centers, arxiv preprint arxiv: , [14] A. Basu, J. Gandhi, J. Chang, M. D. Hill, and M. M. Swift, Efficient virtual memory for big memory servers, in International Symposium on Computer Architecture, 2013, pp [15] J.-S. Kim, X. Qin, and Y. Hsu, Memory characterization of a parallel data mining workload, in Workload Characterization: Methodology and Case Studies, IEEE, 1999, pp [16] M. Karlsson, K. Moore, E. Hagersten, and D. Wood, Memory characterization of the ecperf benchmark, in Workshop on Memory Performance Issues, [17] Y. Xu, E. Frachtenberg, S. Jiang, and M. Palecezny, Characterizing facebook s memcached workload, IEEE Internet Computing, vol. 99, p. 1, [18] Y. Xie, Future memory and interconnect technologies, in Design, Automation & Test in Europe Conference & Exhibition (DATE), IEEE, 2013, pp [19] J. Chen, R. C. Chiang, H. H. Huang, and G. Venkataramani, Energy-aware writes to non-volatile main memory, ACM SIGOPS Operating Systems Review, vol. 45, no. 3, pp , [20] E. Chen et al., Advances and future prospects of spintransfer torque random access memory, Magnetics, IEEE Transactions on, vol. 46, no. 6, pp , [21] S. Huang, J. Huang, J. Dai, T. Xie, and B. Huang, The hibench benchmark suite: Characterization of the mapreducebased data analysis, in International Conference on Data Engineering Workshops. IEEE, 2010, pp [22] M. Stonebraker, Sql databases v. nosql databases, Commun. ACM, vol. 53, no. 4, pp , Apr [23] A. E. Gattiker, F. H. Gebara, A. Gheith, H. P. Hofstee, D. A. Jamsek, J. Li, E. Speight, J. W. Shi, G. C. Chen, and P. W. Wong, Understanding system and architecture for big data, IBM research, [24] J. Chang, K. T. Lim, J. Byrne, L. Ramirez, and P. Ranganathan, Workload diversity and dynamics in big data analytics: implications to system designers, in Workshop on Architectures and Systems for Big Data, 2012, pp [25] H. Yang, Z. Luan, W. Li, D. Qian, and G. Guan, Statisticsbased workload modeling for mapreduce, in Parallel and Distributed Processing Symposium Workshops PhD Forum, 2012, pp [26] B. Atikoglu, Y. Xu, E. Frachtenberg, S. Jiang, and M. Paleczny, Workload analysis of a large-scale key-value store, in ACM International Conference on Measurement and Modeling of Computer Systems, 2012, pp [27] Z. Ren, X. Xu, J. Wan, W. Shi, and M. Zhou, Workload characterization on a production hadoop cluster: A case study on taobao, in IEEE International Symposium on Workload Characterization, 2012, pp [28] Hadoop and memcached: Performance and power characterization and analysis, Journal of Cloud Computing, vol. 1, no. 1,

Benchmarking Cassandra on Violin

Benchmarking Cassandra on Violin Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract

More information

How To Handle Big Data With A Data Scientist

How To Handle Big Data With A Data Scientist III Big Data Technologies Today, new technologies make it possible to realize value from Big Data. Big data technologies can replace highly customized, expensive legacy systems with a standard solution

More information

Evaluating Task Scheduling in Hadoop-based Cloud Systems

Evaluating Task Scheduling in Hadoop-based Cloud Systems 2013 IEEE International Conference on Big Data Evaluating Task Scheduling in Hadoop-based Cloud Systems Shengyuan Liu, Jungang Xu College of Computer and Control Engineering University of Chinese Academy

More information

Big Fast Data Hadoop acceleration with Flash. June 2013

Big Fast Data Hadoop acceleration with Flash. June 2013 Big Fast Data Hadoop acceleration with Flash June 2013 Agenda The Big Data Problem What is Hadoop Hadoop and Flash The Nytro Solution Test Results The Big Data Problem Big Data Output Facebook Traditional

More information

BPOE Research Highlights

BPOE Research Highlights BPOE Research Highlights Jianfeng Zhan ICT, Chinese Academy of Sciences 2013-10- 9 http://prof.ict.ac.cn/jfzhan INSTITUTE OF COMPUTING TECHNOLOGY What is BPOE workshop? B: Big Data Benchmarks PO: Performance

More information

HiBench Introduction. Carson Wang ([email protected]) Software & Services Group

HiBench Introduction. Carson Wang (carson.wang@intel.com) Software & Services Group HiBench Introduction Carson Wang ([email protected]) Agenda Background Workloads Configurations Benchmark Report Tuning Guide Background WHY Why we need big data benchmarking systems? WHAT What is

More information

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications

More information

Energy Efficient MapReduce

Energy Efficient MapReduce Energy Efficient MapReduce Motivation: Energy consumption is an important aspect of datacenters efficiency, the total power consumption in the united states has doubled from 2000 to 2005, representing

More information

HP ProLiant Gen8 vs Gen9 Server Blades on Data Warehouse Workloads

HP ProLiant Gen8 vs Gen9 Server Blades on Data Warehouse Workloads HP ProLiant Gen8 vs Gen9 Server Blades on Data Warehouse Workloads Gen9 Servers give more performance per dollar for your investment. Executive Summary Information Technology (IT) organizations face increasing

More information

In-Memory Databases Algorithms and Data Structures on Modern Hardware. Martin Faust David Schwalb Jens Krüger Jürgen Müller

In-Memory Databases Algorithms and Data Structures on Modern Hardware. Martin Faust David Schwalb Jens Krüger Jürgen Müller In-Memory Databases Algorithms and Data Structures on Modern Hardware Martin Faust David Schwalb Jens Krüger Jürgen Müller The Free Lunch Is Over 2 Number of transistors per CPU increases Clock frequency

More information

BigDataBench. Khushbu Agarwal

BigDataBench. Khushbu Agarwal BigDataBench Khushbu Agarwal Last Updated: May 23, 2014 CONTENTS Contents 1 What is BigDataBench? [1] 1 1.1 SUMMARY.................................. 1 1.2 METHODOLOGY.............................. 1 2

More information

Analysis of Memory Sensitive SPEC CPU2006 Integer Benchmarks for Big Data Benchmarking

Analysis of Memory Sensitive SPEC CPU2006 Integer Benchmarks for Big Data Benchmarking Analysis of Memory Sensitive SPEC CPU2006 Integer Benchmarks for Big Data Benchmarking Kathlene Hurt and Eugene John Department of Electrical and Computer Engineering University of Texas at San Antonio

More information

A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique

A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique A Novel Way of Deduplication Approach for Cloud Backup Services Using Block Index Caching Technique Jyoti Malhotra 1,Priya Ghyare 2 Associate Professor, Dept. of Information Technology, MIT College of

More information

CLOUDDMSS: CLOUD-BASED DISTRIBUTED MULTIMEDIA STREAMING SERVICE SYSTEM FOR HETEROGENEOUS DEVICES

CLOUDDMSS: CLOUD-BASED DISTRIBUTED MULTIMEDIA STREAMING SERVICE SYSTEM FOR HETEROGENEOUS DEVICES CLOUDDMSS: CLOUD-BASED DISTRIBUTED MULTIMEDIA STREAMING SERVICE SYSTEM FOR HETEROGENEOUS DEVICES 1 MYOUNGJIN KIM, 2 CUI YUN, 3 SEUNGHO HAN, 4 HANKU LEE 1,2,3,4 Department of Internet & Multimedia Engineering,

More information

The Flash-Transformed Financial Data Center. Jean S. Bozman Enterprise Solutions Manager, Enterprise Storage Solutions Corporation August 6, 2014

The Flash-Transformed Financial Data Center. Jean S. Bozman Enterprise Solutions Manager, Enterprise Storage Solutions Corporation August 6, 2014 The Flash-Transformed Financial Data Center Jean S. Bozman Enterprise Solutions Manager, Enterprise Storage Solutions Corporation August 6, 2014 Forward-Looking Statements During our meeting today we will

More information

Can Flash help you ride the Big Data Wave? Steve Fingerhut Vice President, Marketing Enterprise Storage Solutions Corporation

Can Flash help you ride the Big Data Wave? Steve Fingerhut Vice President, Marketing Enterprise Storage Solutions Corporation Can Flash help you ride the Big Data Wave? Steve Fingerhut Vice President, Marketing Enterprise Storage Solutions Corporation Forward-Looking Statements During our meeting today we may make forward-looking

More information

An Efficient Hybrid P2P MMOG Cloud Architecture for Dynamic Load Management. Ginhung Wang, Kuochen Wang

An Efficient Hybrid P2P MMOG Cloud Architecture for Dynamic Load Management. Ginhung Wang, Kuochen Wang 1 An Efficient Hybrid MMOG Cloud Architecture for Dynamic Load Management Ginhung Wang, Kuochen Wang Abstract- In recent years, massively multiplayer online games (MMOGs) become more and more popular.

More information

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next

More information

Maximum performance, minimal risk for data warehousing

Maximum performance, minimal risk for data warehousing SYSTEM X SERVERS SOLUTION BRIEF Maximum performance, minimal risk for data warehousing Microsoft Data Warehouse Fast Track for SQL Server 2014 on System x3850 X6 (95TB) The rapid growth of technology has

More information

Evaluating HDFS I/O Performance on Virtualized Systems

Evaluating HDFS I/O Performance on Virtualized Systems Evaluating HDFS I/O Performance on Virtualized Systems Xin Tang [email protected] University of Wisconsin-Madison Department of Computer Sciences Abstract Hadoop as a Service (HaaS) has received increasing

More information

Can High-Performance Interconnects Benefit Memcached and Hadoop?

Can High-Performance Interconnects Benefit Memcached and Hadoop? Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,

More information

A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster

A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster , pp.11-20 http://dx.doi.org/10.14257/ ijgdc.2014.7.2.02 A Load Balancing Algorithm based on the Variation Trend of Entropy in Homogeneous Cluster Kehe Wu 1, Long Chen 2, Shichao Ye 2 and Yi Li 2 1 Beijing

More information

Maximizing Hadoop Performance and Storage Capacity with AltraHD TM

Maximizing Hadoop Performance and Storage Capacity with AltraHD TM Maximizing Hadoop Performance and Storage Capacity with AltraHD TM Executive Summary The explosion of internet data, driven in large part by the growth of more and more powerful mobile devices, has created

More information

Benchmarking Hadoop & HBase on Violin

Benchmarking Hadoop & HBase on Violin Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages

More information

Introducing EEMBC Cloud and Big Data Server Benchmarks

Introducing EEMBC Cloud and Big Data Server Benchmarks Introducing EEMBC Cloud and Big Data Server Benchmarks Quick Background: Industry-Standard Benchmarks for the Embedded Industry EEMBC formed in 1997 as non-profit consortium Defining and developing application-specific

More information

EMC VFCACHE ACCELERATES ORACLE

EMC VFCACHE ACCELERATES ORACLE White Paper EMC VFCACHE ACCELERATES ORACLE VFCache extends Flash to the server FAST Suite automates storage placement in the array VNX protects data EMC Solutions Group Abstract This white paper describes

More information

Surfing the Data Tsunami: A New Paradigm for Big Data Processing and Analytics

Surfing the Data Tsunami: A New Paradigm for Big Data Processing and Analytics Surfing the Data Tsunami: A New Paradigm for Big Data Processing and Analytics Dr. Liangxiu Han Future Networks and Distributed Systems Group (FUNDS) School of Computing, Mathematics and Digital Technology,

More information

Can the Elephants Handle the NoSQL Onslaught?

Can the Elephants Handle the NoSQL Onslaught? Can the Elephants Handle the NoSQL Onslaught? Avrilia Floratou, Nikhil Teletia David J. DeWitt, Jignesh M. Patel, Donghui Zhang University of Wisconsin-Madison Microsoft Jim Gray Systems Lab Presented

More information

Text Mining Approach for Big Data Analysis Using Clustering and Classification Methodologies

Text Mining Approach for Big Data Analysis Using Clustering and Classification Methodologies Text Mining Approach for Big Data Analysis Using Clustering and Classification Methodologies Somesh S Chavadi 1, Dr. Asha T 2 1 PG Student, 2 Professor, Department of Computer Science and Engineering,

More information

Accelerating Server Storage Performance on Lenovo ThinkServer

Accelerating Server Storage Performance on Lenovo ThinkServer Accelerating Server Storage Performance on Lenovo ThinkServer Lenovo Enterprise Product Group April 214 Copyright Lenovo 214 LENOVO PROVIDES THIS PUBLICATION AS IS WITHOUT WARRANTY OF ANY KIND, EITHER

More information

Big Data Performance Growth on the Rise

Big Data Performance Growth on the Rise Impact of Big Data growth On Transparent Computing Michael A. Greene Intel Vice President, Software and Services Group, General Manager, System Technologies and Optimization 1 Transparent Computing (TC)

More information

Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging

Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging Outline High Performance Computing (HPC) Towards exascale computing: a brief history Challenges in the exascale era Big Data meets HPC Some facts about Big Data Technologies HPC and Big Data converging

More information

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Built up on Cisco s big data common platform architecture (CPA), a

More information

Big Data. Value, use cases and architectures. Petar Torre Lead Architect Service Provider Group. Dubrovnik, Croatia, South East Europe 20-22 May, 2013

Big Data. Value, use cases and architectures. Petar Torre Lead Architect Service Provider Group. Dubrovnik, Croatia, South East Europe 20-22 May, 2013 Dubrovnik, Croatia, South East Europe 20-22 May, 2013 Big Data Value, use cases and architectures Petar Torre Lead Architect Service Provider Group 2011 2013 Cisco and/or its affiliates. All rights reserved.

More information

An Approach to Implement Map Reduce with NoSQL Databases

An Approach to Implement Map Reduce with NoSQL Databases www.ijecs.in International Journal Of Engineering And Computer Science ISSN: 2319-7242 Volume 4 Issue 8 Aug 2015, Page No. 13635-13639 An Approach to Implement Map Reduce with NoSQL Databases Ashutosh

More information

Data Refinery with Big Data Aspects

Data Refinery with Big Data Aspects International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 7 (2013), pp. 655-662 International Research Publications House http://www. irphouse.com /ijict.htm Data

More information

Concept of Cache in web proxies

Concept of Cache in web proxies Concept of Cache in web proxies Chan Kit Wai and Somasundaram Meiyappan 1. Introduction Caching is an effective performance enhancing technique that has been used in computer systems for decades. However,

More information

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies

More information

Enterprise Applications

Enterprise Applications Enterprise Applications Chi Ho Yue Sorav Bansal Shivnath Babu Amin Firoozshahian EE392C Emerging Applications Study Spring 2003 Functionality Online Transaction Processing (OLTP) Users/apps interacting

More information

Networking in the Hadoop Cluster

Networking in the Hadoop Cluster Hadoop and other distributed systems are increasingly the solution of choice for next generation data volumes. A high capacity, any to any, easily manageable networking layer is critical for peak Hadoop

More information

Exploring RAID Configurations

Exploring RAID Configurations Exploring RAID Configurations J. Ryan Fishel Florida State University August 6, 2008 Abstract To address the limits of today s slow mechanical disks, we explored a number of data layouts to improve RAID

More information

Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture

Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture Analysis and Optimization of Massive Data Processing on High Performance Computing Architecture He Huang, Shanshan Li, Xiaodong Yi, Feng Zhang, Xiangke Liao and Pan Dong School of Computer Science National

More information

How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time

How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time SCALEOUT SOFTWARE How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time by Dr. William Bain and Dr. Mikhail Sobolev, ScaleOut Software, Inc. 2012 ScaleOut Software, Inc. 12/27/2012 T wenty-first

More information

Efficient Data Replication Scheme based on Hadoop Distributed File System

Efficient Data Replication Scheme based on Hadoop Distributed File System , pp. 177-186 http://dx.doi.org/10.14257/ijseia.2015.9.12.16 Efficient Data Replication Scheme based on Hadoop Distributed File System Jungha Lee 1, Jaehwa Chung 2 and Daewon Lee 3* 1 Division of Supercomputing,

More information

Fast, Low-Overhead Encryption for Apache Hadoop*

Fast, Low-Overhead Encryption for Apache Hadoop* Fast, Low-Overhead Encryption for Apache Hadoop* Solution Brief Intel Xeon Processors Intel Advanced Encryption Standard New Instructions (Intel AES-NI) The Intel Distribution for Apache Hadoop* software

More information

Log Mining Based on Hadoop s Map and Reduce Technique

Log Mining Based on Hadoop s Map and Reduce Technique Log Mining Based on Hadoop s Map and Reduce Technique ABSTRACT: Anuja Pandit Department of Computer Science, [email protected] Amruta Deshpande Department of Computer Science, [email protected]

More information

Oracle Database In-Memory The Next Big Thing

Oracle Database In-Memory The Next Big Thing Oracle Database In-Memory The Next Big Thing Maria Colgan Master Product Manager #DBIM12c Why is Oracle do this Oracle Database In-Memory Goals Real Time Analytics Accelerate Mixed Workload OLTP No Changes

More information

Using Synology SSD Technology to Enhance System Performance Synology Inc.

Using Synology SSD Technology to Enhance System Performance Synology Inc. Using Synology SSD Technology to Enhance System Performance Synology Inc. Synology_SSD_Cache_WP_ 20140512 Table of Contents Chapter 1: Enterprise Challenges and SSD Cache as Solution Enterprise Challenges...

More information

Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, SIGMOD, New York, June 2013

Petabyte Scale Data at Facebook. Dhruba Borthakur, Engineer at Facebook, SIGMOD, New York, June 2013 Petabyte Scale Data at Facebook Dhruba Borthakur, Engineer at Facebook, SIGMOD, New York, June 2013 Agenda 1 Types of Data 2 Data Model and API for Facebook Graph Data 3 SLTP (Semi-OLTP) and Analytics

More information

Keywords: Big Data, HDFS, Map Reduce, Hadoop

Keywords: Big Data, HDFS, Map Reduce, Hadoop Volume 5, Issue 7, July 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Configuration Tuning

More information

Oracle Big Data SQL Technical Update

Oracle Big Data SQL Technical Update Oracle Big Data SQL Technical Update Jean-Pierre Dijcks Oracle Redwood City, CA, USA Keywords: Big Data, Hadoop, NoSQL Databases, Relational Databases, SQL, Security, Performance Introduction This technical

More information

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack

Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack HIGHLIGHTS Real-Time Results Elasticsearch on Cisco UCS enables a deeper

More information

The Comprehensive Performance Rating for Hadoop Clusters on Cloud Computing Platform

The Comprehensive Performance Rating for Hadoop Clusters on Cloud Computing Platform The Comprehensive Performance Rating for Hadoop Clusters on Cloud Computing Platform Fong-Hao Liu, Ya-Ruei Liou, Hsiang-Fu Lo, Ko-Chin Chang, and Wei-Tsong Lee Abstract Virtualization platform solutions

More information

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM Sneha D.Borkar 1, Prof.Chaitali S.Surtakar 2 Student of B.E., Information Technology, J.D.I.E.T, [email protected] Assistant Professor, Information

More information

Detection of Distributed Denial of Service Attack with Hadoop on Live Network

Detection of Distributed Denial of Service Attack with Hadoop on Live Network Detection of Distributed Denial of Service Attack with Hadoop on Live Network Suchita Korad 1, Shubhada Kadam 2, Prajakta Deore 3, Madhuri Jadhav 4, Prof.Rahul Patil 5 Students, Dept. of Computer, PCCOE,

More information

CloudRank-D:A Benchmark Suite for Private Cloud Systems

CloudRank-D:A Benchmark Suite for Private Cloud Systems CloudRank-D:A Benchmark Suite for Private Cloud Systems Jing Quan Institute of Computing Technology, Chinese Academy of Sciences and University of Science and Technology of China HVC tutorial in conjunction

More information

Big Data Storage Architecture Design in Cloud Computing

Big Data Storage Architecture Design in Cloud Computing Big Data Storage Architecture Design in Cloud Computing Xuebin Chen 1, Shi Wang 1( ), Yanyan Dong 1, and Xu Wang 2 1 College of Science, North China University of Science and Technology, Tangshan, Hebei,

More information

Reconfigurable Architecture Requirements for Co-Designed Virtual Machines

Reconfigurable Architecture Requirements for Co-Designed Virtual Machines Reconfigurable Architecture Requirements for Co-Designed Virtual Machines Kenneth B. Kent University of New Brunswick Faculty of Computer Science Fredericton, New Brunswick, Canada [email protected] Micaela Serra

More information

Map-Parallel Scheduling (mps) using Hadoop environment for job scheduler and time span for Multicore Processors

Map-Parallel Scheduling (mps) using Hadoop environment for job scheduler and time span for Multicore Processors Map-Parallel Scheduling (mps) using Hadoop environment for job scheduler and time span for Sudarsanam P Abstract G. Singaravel Parallel computing is an base mechanism for data process with scheduling task,

More information

Introduction. Various user groups requiring Hadoop, each with its own diverse needs, include:

Introduction. Various user groups requiring Hadoop, each with its own diverse needs, include: Introduction BIG DATA is a term that s been buzzing around a lot lately, and its use is a trend that s been increasing at a steady pace over the past few years. It s quite likely you ve also encountered

More information

Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging

Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging In some markets and scenarios where competitive advantage is all about speed, speed is measured in micro- and even nano-seconds.

More information

Scalable Multiple NameNodes Hadoop Cloud Storage System

Scalable Multiple NameNodes Hadoop Cloud Storage System Vol.8, No.1 (2015), pp.105-110 http://dx.doi.org/10.14257/ijdta.2015.8.1.12 Scalable Multiple NameNodes Hadoop Cloud Storage System Kun Bi 1 and Dezhi Han 1,2 1 College of Information Engineering, Shanghai

More information

The Impact of Memory Subsystem Resource Sharing on Datacenter Applications. Lingia Tang Jason Mars Neil Vachharajani Robert Hundt Mary Lou Soffa

The Impact of Memory Subsystem Resource Sharing on Datacenter Applications. Lingia Tang Jason Mars Neil Vachharajani Robert Hundt Mary Lou Soffa The Impact of Memory Subsystem Resource Sharing on Datacenter Applications Lingia Tang Jason Mars Neil Vachharajani Robert Hundt Mary Lou Soffa Introduction Problem Recent studies into the effects of memory

More information

BIG DATA TECHNOLOGY. Hadoop Ecosystem

BIG DATA TECHNOLOGY. Hadoop Ecosystem BIG DATA TECHNOLOGY Hadoop Ecosystem Agenda Background What is Big Data Solution Objective Introduction to Hadoop Hadoop Ecosystem Hybrid EDW Model Predictive Analysis using Hadoop Conclusion What is Big

More information

Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control

Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control EP/K006487/1 UK PI: Prof Gareth Taylor (BU) China PI: Prof Yong-Hua Song (THU) Consortium UK Members: Brunel University

More information

Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies

Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at: www.ijarcsms.com Image

More information

BiDAl: Big Data Analyzer for Cluster Traces

BiDAl: Big Data Analyzer for Cluster Traces BiDAl: Big Data Analyzer for Cluster Traces Alkida Balliu, Dennis Olivetti, Ozalp Babaoglu, Moreno Marzolla, Alina Sirbu Department of Computer Science and Engineering University of Bologna, Italy BigSys

More information

Telecom Data processing and analysis based on Hadoop

Telecom Data processing and analysis based on Hadoop COMPUTER MODELLING & NEW TECHNOLOGIES 214 18(12B) 658-664 Abstract Telecom Data processing and analysis based on Hadoop Guofan Lu, Qingnian Zhang *, Zhao Chen Wuhan University of Technology, Wuhan 4363,China

More information

Infrastructure Matters: POWER8 vs. Xeon x86

Infrastructure Matters: POWER8 vs. Xeon x86 Advisory Infrastructure Matters: POWER8 vs. Xeon x86 Executive Summary This report compares IBM s new POWER8-based scale-out Power System to Intel E5 v2 x86- based scale-out systems. A follow-on report

More information

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database Managing Big Data with Hadoop & Vertica A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database Copyright Vertica Systems, Inc. October 2009 Cloudera and Vertica

More information

An Oracle White Paper June 2012. High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database

An Oracle White Paper June 2012. High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database An Oracle White Paper June 2012 High Performance Connectors for Load and Access of Data from Hadoop to Oracle Database Executive Overview... 1 Introduction... 1 Oracle Loader for Hadoop... 2 Oracle Direct

More information

Performance Comparison of SQL based Big Data Analytics with Lustre and HDFS file systems

Performance Comparison of SQL based Big Data Analytics with Lustre and HDFS file systems Performance Comparison of SQL based Big Data Analytics with Lustre and HDFS file systems Rekha Singhal and Gabriele Pacciucci * Other names and brands may be claimed as the property of others. Lustre File

More information

Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing

Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing Shareability and Locality Aware Scheduling Algorithm in Hadoop for Mobile Cloud Computing Hsin-Wen Wei 1,2, Che-Wei Hsu 2, Tin-Yu Wu 3, Wei-Tsong Lee 1 1 Department of Electrical Engineering, Tamkang University

More information

VirtualCenter Database Performance for Microsoft SQL Server 2005 VirtualCenter 2.5

VirtualCenter Database Performance for Microsoft SQL Server 2005 VirtualCenter 2.5 Performance Study VirtualCenter Database Performance for Microsoft SQL Server 2005 VirtualCenter 2.5 VMware VirtualCenter uses a database to store metadata on the state of a VMware Infrastructure environment.

More information

An efficient Join-Engine to the SQL query based on Hive with Hbase Zhao zhi-cheng & Jiang Yi

An efficient Join-Engine to the SQL query based on Hive with Hbase Zhao zhi-cheng & Jiang Yi International Conference on Applied Science and Engineering Innovation (ASEI 2015) An efficient Join-Engine to the SQL query based on Hive with Hbase Zhao zhi-cheng & Jiang Yi Institute of Computer Forensics,

More information

Impact of Big Data: Networking Considerations and Case Study

Impact of Big Data: Networking Considerations and Case Study 30 Impact of Big Data: Networking Considerations and Case Study Yong-Hee Jeon Catholic University of Daegu, Gyeongsan, Rep. of Korea Summary which exceeds the range possible to store, manage, and Due to

More information

Dynamic resource management for energy saving in the cloud computing environment

Dynamic resource management for energy saving in the cloud computing environment Dynamic resource management for energy saving in the cloud computing environment Liang-Teh Lee, Kang-Yuan Liu, and Hui-Yang Huang Department of Computer Science and Engineering, Tatung University, Taiwan

More information

The MAX5 Advantage: Clients Benefit running Microsoft SQL Server Data Warehouse (Workloads) on IBM BladeCenter HX5 with IBM MAX5.

The MAX5 Advantage: Clients Benefit running Microsoft SQL Server Data Warehouse (Workloads) on IBM BladeCenter HX5 with IBM MAX5. Performance benefit of MAX5 for databases The MAX5 Advantage: Clients Benefit running Microsoft SQL Server Data Warehouse (Workloads) on IBM BladeCenter HX5 with IBM MAX5 Vinay Kulkarni Kent Swalin IBM

More information

Binary search tree with SIMD bandwidth optimization using SSE

Binary search tree with SIMD bandwidth optimization using SSE Binary search tree with SIMD bandwidth optimization using SSE Bowen Zhang, Xinwei Li 1.ABSTRACT In-memory tree structured index search is a fundamental database operation. Modern processors provide tremendous

More information

System Requirements Table of contents

System Requirements Table of contents Table of contents 1 Introduction... 2 2 Knoa Agent... 2 2.1 System Requirements...2 2.2 Environment Requirements...4 3 Knoa Server Architecture...4 3.1 Knoa Server Components... 4 3.2 Server Hardware Setup...5

More information

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance

Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance Intel Solid- State Drive Data Center P3700 Series NVMe Hybrid Storage Performance Hybrid Storage Performance Gains for IOPS and Bandwidth Utilizing Colfax Servers and Enmotus FuzeDrive Software NVMe Hybrid

More information

ESS event: Big Data in Official Statistics. Antonino Virgillito, Istat

ESS event: Big Data in Official Statistics. Antonino Virgillito, Istat ESS event: Big Data in Official Statistics Antonino Virgillito, Istat v erbi v is 1 About me Head of Unit Web and BI Technologies, IT Directorate of Istat Project manager and technical coordinator of Web

More information

The Methodology Behind the Dell SQL Server Advisor Tool

The Methodology Behind the Dell SQL Server Advisor Tool The Methodology Behind the Dell SQL Server Advisor Tool Database Solutions Engineering By Phani MV Dell Product Group October 2009 Executive Summary The Dell SQL Server Advisor is intended to perform capacity

More information

Performance Impacts of Non-blocking Caches in Out-of-order Processors

Performance Impacts of Non-blocking Caches in Out-of-order Processors Performance Impacts of Non-blocking Caches in Out-of-order Processors Sheng Li; Ke Chen; Jay B. Brockman; Norman P. Jouppi HP Laboratories HPL-2011-65 Keyword(s): Non-blocking cache; MSHR; Out-of-order

More information

Microsoft SQL Server 2014 Fast Track

Microsoft SQL Server 2014 Fast Track Microsoft SQL Server 2014 Fast Track 34-TB Certified Data Warehouse 103-TB Maximum User Data Tegile Systems Solution Review 2U Design: Featuring Tegile T3800 All-Flash Storage Array http:// www.tegile.com/solutiuons/sql

More information

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC

More information

2009 Oracle Corporation 1

2009 Oracle Corporation 1 The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material,

More information

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren

News and trends in Data Warehouse Automation, Big Data and BI. Johan Hendrickx & Dirk Vermeiren News and trends in Data Warehouse Automation, Big Data and BI Johan Hendrickx & Dirk Vermeiren Extreme Agility from Source to Analysis DWH Appliances & DWH Automation Typical Architecture 3 What Business

More information

Big Data Are You Ready? Thomas Kyte http://asktom.oracle.com

Big Data Are You Ready? Thomas Kyte http://asktom.oracle.com Big Data Are You Ready? Thomas Kyte http://asktom.oracle.com The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated

More information

Big Data: Study in Structured and Unstructured Data

Big Data: Study in Structured and Unstructured Data Big Data: Study in Structured and Unstructured Data Motashim Rasool 1, Wasim Khan 2 [email protected], [email protected] Abstract With the overlay of digital world, Information is available

More information

Accelerating Business Intelligence with Large-Scale System Memory

Accelerating Business Intelligence with Large-Scale System Memory Accelerating Business Intelligence with Large-Scale System Memory A Proof of Concept by Intel, Samsung, and SAP Executive Summary Real-time business intelligence (BI) plays a vital role in driving competitiveness

More information

Analytics in the Cloud. Peter Sirota, GM Elastic MapReduce

Analytics in the Cloud. Peter Sirota, GM Elastic MapReduce Analytics in the Cloud Peter Sirota, GM Elastic MapReduce Data-Driven Decision Making Data is the new raw material for any business on par with capital, people, and labor. What is Big Data? Terabytes of

More information