SERVER DESIGNS FOR WAREHOUSE- COMPUTING ENVIRONMENTS
|
|
|
- Hector Peters
- 10 years ago
- Views:
Transcription
1 [3B2-3] mmi d 3/2/09 14:56 Page SERVER DESIGNS FOR WAREHOUSE- COMPUTING ENVIRONMENTS... THE ENORMOUS SCALE OF WAREHOUSE-COMPUTING ENVIRONMENTS LEADS TO UNIQUE REQUIREMENTS IN WHICH COST AND POWER FIGURE PROMINENTLY. MODELS AND METRICS QUANTIFYING THESE REQUIREMENTS, ALONG WITH A BENCHMARK SUITE TO CAPTURE WORKLOAD BEHAVIOR, HELP IDENTIFY BOTTLENECKS AND EVALUATE SOLU- TIONS. AHOLISTIC APPROACH LEADS TO A NEW SYSTEM ARCHITECTURE INCORPORATING VOLUME NON-SERVER-CLASS COMPONENTS IN NOVEL PACKAGING SOLUTIONS, WITH MEMORY SHARING AND FLASH-BASED DISK CACHING. Kevin Lim University of Michigan Parthasarathy Ranganathan Jichuan Chang Chandrakant Patel Hewlett-Packard Labs Trevor Mudge Steven K. Reinhardt University of Michigan...In recent years, one of the biggest trends in the server market has been the emergenceofthelarge-scaledatacenter, driven by Internet-sector growth. Indeed, recent market research identifies the Internet sector as the fastest-growing segment of the overall server market, growing by to 65 percent every year, and accounting for more than 65 percent of low-end-server revenue growth in Furthermore, several recent news articles and keynote speeches have highlighted this area s importance. 1,2 One of the most interesting aspects of this growth is the unique set of challenges it presents for server design. Internet-sector infrastructures have millions of users running on hundreds of thousands of servers, making the ability to scale-out server configurations a key design requirement. Experts have compared these environments to a large warehouse-style computer, with distributed applications such as mail, search, and so on. 1 For companies in this market, data-center infrastructure including power and cooling can be the largest capital and operating expense, motivating /09/$25.00 c 2009 IEEE Published by the IEEE Computer Society companies to focus on the sweet spot of commodity pricing and energy efficiency. Furthermore, the high volume of this market and the dominance of a few key players for example, Google, Microsoft, Amazon allow for exploring novel approaches such as custom-designed servers in greenfield data centers, newly built from the ground up. Our work addresses these challenges and opportunities by developing workloads and metrics targeted toward large-scale data centers, and using these tools to build and analyze new system architectures optimized for the Internet sector. We created an evaluation infrastructure including a benchmark suite for warehouse-computing workloads, and detailed performance, cost, and power models and metrics. These tools let us see the contributions of all layers within large-scale data centers, and lead us to the identification of the, cooling, memory, and disk as the key performance-per-cost bottlenecks. Using our understanding of the importance of these components, we created a system architecture that holistically addresses each... 41
2 [3B2-3] mmi d 3/2/09 14:56 Page TOP PICKS Table 1. Details of our benchmark suite representing Internet-sector workloads. Workload Emphasis Description Performance metric Web search Web mail Video hosting MapReduce Role of unstructured data Interactive Internet services Use of rich media Web as a platform Nutch running on Apache Tomcat with a 1.3-Gbyte index (1.3 million documents) and 25 percent of index terms cached in memory, with a 2-Gbyte Java heap size. Quality of service (QoS) requires more than 95 percent of queries to take under 0.5 seconds. Front end: Squirrelmail with Apache and PHP4. Back end: Courier-Internet Message Access Protocol and Exim. One-thousand virtual users with 7 Gbytes of s. and attachment sizes and use patterns modeled after MS Exchange 2003 LoadSim for heavy users. QoS requires more than 95 percent requests to take under 0.8 second. Modified SPECweb2005 Support workload with YouTube traffic characteristics. Apache Tomcat with Rock httpd server. Apache Hadoop with four threads per and 1.5-Gbyte Java heap size. Two workloads: distributed file write (mapred-wr) and word count (mapred-wc). Request per second (RPS) with QoS RPS with QoS RPS with QoS Execution time 42 IEEE MICRO layer through novel features, including the use of low-cost, low-power components from the high-volume, embedded market; new packaging solutions; memory sharing; and flash-based disk caching. Our results are promising, providing an average twofold improvement in performance per dollar for our benchmarks. More importantly, our work highlights the potential of cost-efficient, ensemble-level design, and suggests the usefulness of further research in this area. Evaluation environment Due to the novelty of warehouse-computing environments and the lack of public information available, we developed a new workload suite, as well as cost, power, and performance models tailored to this domain. In addition, we propose a performance-per-cost metric for large-scale data centers, and using this metric along with our tools, we perform an analysis to identify the key cost bottlenecks that need to be addressed using new system architectures. Workloads One key contribution of our work is the development of a benchmark suite that models warehouse-computing workloads. Applications in this area are different from traditional benchmarks due to their scale and level of user involvement. This difference is reflected by our benchmark suite s focus on large data sets, user interactivity, and cluster-level processing. We based our suite on open source software and publicly available real-world traces in the hope that the suite will be useful to other researchers. The benchmark suite comprises four workloads representative of different services in Internet-sector data centers. Our workloads include Web search, Web mail, video hosting, and MapReduce, 3 which emphasize unstructured data, user involvement, rich media types, and the Web as a platform, respectively. Table 1 lists these benchmarks in greater detail, including the software used for each benchmark. To emulate the large data sets typically found in these workloads, we have used a wide array of sources, including data gathered from public sites (Wikipedia search index), internally collected statistics ( statistics from within the University of Michigan), and previously published data (recorded YouTube traffic at an edge network 4 ). To make simulation
3 [3B2-3] mmi d 3/2/09 14:56 Page 43 feasible, we use scaled-down versions of these data sets. One issue adding to the complexity of our workloads is the importance of throughput and latency in the Internet sector, both the number of requests per second processed and the network response time are key performance measures. Thus, to drive our workloads, we have developed software clients that change their request rates dynamically to achieve the highest possible throughput while maintaining quality-of-service guarantees for response time. Note that this does not apply to Map- Reduce, which is a cluster-processing job. We believe this benchmark suite is a good representative of Internet-sector workloads for our study, and provides a sound basis for future expansion. Metrics and models In addition to using these benchmarks to compare the performance of different server designs, we defined several key metrics and models that give a detailed overview of Internet-sector environments. Our primary metric was the sustainable performance divided by total cost of ownership (TCO). For TCO, we assumed a three-year depreciation cycle and consider costs associated with base hardware, burdened power and cooling, and real estate. Our cost model consists of two main components: base hardware and burdened power and cooling costs. For the base hardware costs, we aggregate the costs of the individual components that is,, memory, disk, board, and power and cooling components at a per-server level. We combine these costs at the rack level, and then consider additional switch and enclosure costs for the rack. For the power and cooling costs, we use two subcomponents. We first compute the rack-level power consumption (P_consumed) by summing the power used for the, memory, disk, power and cooling, and the rest of the board at the per-server level, and additional switch power at the rack level. We proratethisbyanactivityfactorof0.75 because the published power numbers we use are worst-case numbers, which are higher than the actual power consumption. We then use P_consumed to determine the burdened Details Table 2. Servers cost models. Board and management Power and fans Total server Switch and rack Servers per rack Three-year power and cooling Total costs Server 1 cost (dollars) 1, ,225 2,750 2,464 5,758 Server 2 cost (dollars) ,620 2,750 1,561 3,249 Table 3. Servers power models. Details Board and management Power and fans Total server Server 1 power (watts) Server 2 power (watts) cost of power and cooling using the methodology discussed by Patel et al., 5 which factors in electricity costs for the system and the cooling, as well as the amortized infrastructure costs for power and cooling. To evaluate performance, we used Hewlett- Packard Labs COTSon simulator, 6 which is based on AMD s SimNow infrastructure. 7 COTSon is a validated full-system x86/x86-64 simulator, running 64-bit Debian Linux with the kernel. COTSon is flexible enough to let us simulate several server configurations and multisystem models. Cost analysis We use these workloads, metrics, and detailed models as tools to evaluate the contributions of all layers in a warehouse-computing environment. Tables 2 and 3 list the costs and JANUARY/FEBRUARY
4 [3B2-3] mmi d 3/2/09 14:56 Page TOP PICKS Board power and cooling 9% power and cooling 2% power and cooling 6% power and cooling 22% Fan power and cooling 8% Rack hardware 2% hardware 20% hardware 11% hardware 4% Board Fan hardware 8% hardware 8% Figure 1. Cost breakdown for server 2, a low-end server-class system configuration. Low-power, lowcost, embedded-class processors offer superior performance/ total-cost-of-ownership results Novel and efficient packaging enables multiple compact servers per blade Remote shared memory blades require less total memory, reducing cost and power Local flash-based disk caching lowers power, boosts performance Figure 2. Our system s architecture illustrating a holistic approach to improving component costs. 44 IEEE MICRO power use for existing midrange (server 1) and low-end (server 2) server configurations. Figure 1 shows a breakdown of the total costs for the low-end server, separated as infrastructure and burdened power and cooling. This data highlights several interesting points. We see that power and cooling costs are comparable to hardware costs. Although the hardware and power and cooling are the two largest total costs, several other components including memory and disk contribute similarly. Thus, to achieve compelling performance-per-dollar advantages, we need solutions to holistically address multiple components. New server architecture Based on the insights gained from the previous cost and power-use analysis, we consider three key issues aimed at a holistic solution: reducing overall -related costs by using lower-cost, lower-power (but also lower-performance) nonserver processors; reducing the burdened costs of power with novel packaging solutions; and improving the memory and disk subsystems using solutions geared toward low power and low aggregate costs. Figure 2 shows our system architecture, which uses novel designs and optimizations to address each of these issues. Low-power s Server designs for databases or highperformance computing have traditionally focused on obtaining the highest performance per server. However, the nature of Internet sector workloads encourages scaleout environments, which use many servers to achieve high performance. This requires instead a focus on performance per cost per server. Indeed, publications by large Internetsector companies, such as Google, 8 exhibit the usefulness of building servers using commodity desktop-pc parts. The logic behind this is that volume drives cost. Compared to servers that have a limited market and higher price margins, commodity PCs have a much larger market that enables lower prices. Additionally, these systems don t include cost premiums for high-end features that the built-in reliability support of the software stack for Internet-sector workloads makes redundant. In this section, we evaluate the effectiveness of low-end servers and desktops for the warehouse-computing market. We take the focus on performance per cost one step further, exploring an alternative commodity market: the embedded and mobile segment. Trends in transistor scaling and embeddedprocessor design have brought powerful,
5 [3B2-3] mmi d 3/2/09 14:56 Page 45 Relative hardware costs (%) (a) Server 1 Server 2 Rack + switch Power + fan Board + management Desktop Mobile Embedded system 1 Embedded system 2 Relative cooling costs (%) (b) Server 1 Server 2 Desktop Mobile Embedded system 1 Embedded system 2 Figure 3. Summary of benefits from using low-cost, low-power s from nonserver markets: infrastructure costs (a) and cooling costs (b). general-purpose, multicore processors to the embedded arena. Devices using embedded s are shipped in even higher volume than desktops, leading to greater cost savings. Additionally, embedded parts are often designed for minimal power consumption due to their use in mobile systems. Power is a large portion of total lifecycle costs, so greater power efficiency reduces costs. The key open question is whether these cost and power benefits can offset the performance degradation relative to the baseline server. To perform this evaluation, we consider six system configurations: server 1 (midrange), server 2 (low end), desktop, mobile, embedded 1 (midrange), and embedded 2 (low end). All the systems are dual core excluding server 1, which is a quad core, and embedded system 2, which is a single core. All have 4 Gbytes of memory. Servers 1 and 2 use fully buffered dual inline memory modules (FB-DIMMs); desktop, mobile, and embedded system 1 use double data rate (DDR) 2 memory; and embedded system 2 uses DDR1 memory. Server 1 has a 15,000-revolutions-per-minute disk and a 10-Gbit network interface controller, while the other systems have a 7,200-RPM disk and a 1-Gbit NIC. Our goal for these experiments is to isolate the effect of the processor type, so we keep memory and disk capacity constant. Figures 3a and 3b summarize the breakdown of infrastructure costs and the burdened power and cooling costs. Table 4 shows the variation in performance per cost for each system. In the table, the harmonic mean row shows the average performance per cost, computed as the harmonic mean of the throughput and reciprocal of execution times divided by total system costs. As Figure 3a illustrates, at a per-system level, the hardware costs are dramatically lower for the consumer systems. The biggest reduction comes in the component. The use of consumer technologies, such as DDR2 memory, leads to reductions in other components as well. The costs of the desktop system are 25 percent of the costs of the server 1 configuration, whereas the costs of embedded system 1 are only 15 percent of the costs of the server 1 configuration. In the mobile system, we see higher costs relative to the desktop because of the premium for low-power components in this market. Figure 3b shows similar trends for power and cooling costs. The desktop system has 60 percent lower power and cooling costs JANUARY/FEBRUARY
6 [3B2-3] mmi d 3/2/09 14:56 Page TOP PICKS Table 4. Normalized performance per total-cost-of-ownership efficiency. Embedded system 1 (%) Embedded system 2 (%) Workload Server 2 (%) Desktop (%) Mobile (%) Web search Web mail Video hosting MapRed-wc MapRed-wr Harmonic mean IEEE MICRO than server 1, but the embedded system 1 system does even better, saving 85 percent of the costs. Table 4 shows significant improvements in performance per cost for the desktop, mobile, and embedded system 1 relative to server 1; however, embedded system 2 doesn t perform well. Given the relative losses in performance to the benefits in costs, these systems often have superior performance per cost compared to server 1. On video hosting and MapReduce, embedded system 1 achieves three- to sixfold improvement in performance per total costs due to these workloads being primarily network or disk bound. Embedded system 1 achieves an improvement of 67 percent for Web search, but has a net degradation in performance per cost for Web mail because of a significant performance decrease. Overall, our workloads show a benefit from lower-end consumer platforms optimized for power and costs, compared to server 1. A key result from our study is that embedded systems could offer more cost savings at the same performance; but the choice of embedded platform is important (for example, embedded system 1 versus embedded system 2). It must be noted that these results hold true for our workloads, but more study is needed before we can generalize these results to all variations of Internetsector workloads. Cooling and packaging Ourdiscussioninthe Costanalysis section notes that after the processor, the next largest cost factor is the burdened cost of cooling. Using lower-power systems presents new opportunities for smaller form factor boards, which in turn allow for cooling system optimizations. Here, we discuss two such optimizations dual-entry enclosures with directed airflow and boardlevel aggregated heat removal using blade servers as the exemplar for the rest of our discussions, because they re well known in the market. Dual-entry enclosures with directed airflow. Figure 2 shows how we can redesign a server-level enclosure to allow inserting blades from front and back to attach to a midplane within the enclosure. The idea behind this is to partition the airflow and allow directing of cold air vertically through the blades. We achieve this by increasing the enclosure s volume to create an inlet and exhaust plenum, and direct the airflow in the directions indicated by the arrows in Figure 2. This design maintains airflow through all the blades in parallel from intake plenum to exhaust plenum. Compared to conventional blade enclosures, which force air directly from front to back, this design has a shorter flow length (the distance traversed by the air), lower preheat (the temperature of the air hitting the blades), and reduced pressure drop and volume flow. Our thermomechanical analysis of the thermal-resistance, airflow improvements with this design shows an almost 50 percent improvement in cooling efficiencies. Compared to the baseline that allows forty 1U servers on each rack, our design allows forty 75-watt blades in a 5U enclosure, allowing 320 systems on each rack. Board-level aggregated heat removal. Figure 4 shows an even more radical packaging design. With low-power systems, we can consider
7 [3B2-3] mmi d 3/2/09 14:56 Page 47 servers with much smaller form factors that are integrated on conventional blades that fit into an enclosure. Figure 4 shows an innovative packaging scheme that aggregates the power-dissipating components at the device and package level. We intersperse the smaller form factor server modules with planar heat pipes that transfer heat to a central location with an effective conductivity three times that of copper. The design removes aggregated heat with a larger, optimized heat sink, which enables channeling airflow through a single heat sink as opposed to multiple separate conventional heat sinks. The increases in conductivity and heat-extraction area lead to more effective cooling. We can connect the smaller blades through different interfaces, such as Computeron-Module Express or Advanced Mezzanine Card interfaces. With higher power budgets, one large blade can support four smaller modules, allowing 1,250 systems on each rack. These cooling optimizations could improve efficiencies two- to fourfold. Although we use specialized designs, our cooling solutions should perform effectively in other enterprise environments. When combined with the significant and growing fraction of the market represented by warehouse-computing environments, these designs should have enough volume to drive commoditization. Other optimizations Our cost analysis showed that memory and disk are other key factors in server systems overall performance per cost. We developed optimized memory and disk subsystems that lower costs and power while maintaining high performance by exploiting memory and disk behaviors and characteristics. Our memory design leverages the observation that memory demands across workloads can vary widely. Studies have shown that perserver sizing for peak loads can lead to significant ensemble-level overprovisioning; 9,10 this problem is exacerbated in large-scale data centers running a variety of applications. To address memory overprovisioning, we provision memory at a coarser, per-blade-chassis granularity, sizing each larger unit to meet the expected aggregate peak demand. In our design, we connect servers within a blade enclosure to a memory blade, which provides Condenser Planar heat pipe Aluminum plate Evaporator Land grid array Figure 4. Aggregated heat removal design. Planar heat pipe a remote memory pool that is dynamically partitioned among the servers. By exploiting locality while using the remote memory, the systems can have a reduced local memory and still maintain high performance. Within asingleenclosure,weconnecttheserver and memory blades via a high-speed interconnect (for example, PCIe). We provide the servers with simple hardware support and make minor changes to the operating system or hypervisor to let applications transparently use the remote memory. 11 This hierarchical design enables further optimizations, such as the use of power-down modes and lower-cost DRAM. By provisioning memory more accurately across servers, our design obtains notable power and cost savings. Our disk optimizations use low-power, laptop-class disks to reduce power consumption, and have flash-based disk caching to provide higher performance. We move the laptop disks to a basic storage area network (SAN) interfaced through the Serial ATA interface. By using a SAN, we can make individual server blades smaller, allowing small module form factors. Although the disks have lower power and a smaller form factor, they come with the tradeoffs of lower performance and higher price. To offset this, we use a flash-based disk-caching mechanism, 12 with the flash located on the server board itself. The flash holds any recently JANUARY/FEBRUARY
8 [3B2-3] mmi d 3/2/09 14:56 Page TOP PICKS Performance/total-cost-of-ownership improvement (%) Web search Web mail Video hosting MapReduce-wc MapReduce-wr Harmonic mean Figure 5. Performance per total-cost-of-ownership improvements from our unified design compared to a baseline midrange server. 48 IEEE MICRO accessed pages from the disk, and the system checks its contents any time a page is not found in the operating system s page cache. Flash-based disk caching combined with lowpower laptop disks provides performanceper-cost advantages over traditional serverclass disks and allows even greater compaction. Unified design Our solutions so far have targeted the top performance-per-cost factors we identified in current servers:, cooling and packaging, memory, and disk. Although we ve discussed their benefits in isolation, we designed all of our optimizations to work in combination, addressing our holistic design motivation. Figure 2 shows our unified design, which uses embedded-class, dual-core processors and small form-factor blades with aggregated cooling, housed in an enclosure with directed airflow. We use a remote memory blade and remote low-power disks with flashbased disk caching to enable this level of compaction. The use of the low-power components and efficient cooling and packaging are critical to making this super-dense server feasible for large-scale data centers. Although some of the changes required for this configuration assume custom components, they ll likely become cost-effective in a few years, leveraging the volume in this market. Figure 5 shows how our solution significantly improves cost and power efficiencies compared to the baseline, midrange server system. For the video hosting and Map- Reduce benchmarks, the performance and TCO improves by 3.5 to 6 times for the next-generation solution. These benefits come equally from infrastructure costs and power savings. Web search gets lower benefits 10 to 70 percent improvement and Web mail sees degradations (20 percent for the next-generation system). Figure 5 also shows the harmonic mean across our benchmarks. Overall, our solution can improve sustained throughput per total infrastructure dollar by 2 times. That is, for the same performance as the baseline, our solution gets a 60 percent reduction in power, a 55 percent reduction in overall costs, and requires 30 percent less rack space (assuming four air-cooled embedded blades on each larger blade). We don t intend for our proposed techniques to be the final word on warehouse-computing designs, but rather to illustrate the substantial improvements achievable when system architects take an ensemblelevel view. Overall, as enterprises gravitate toward ever more cost-conscious and datacenter-level solutions, holistic approaches such astheonesusedinthisarticlewilllikelybea key part of future system designs. MICRO References 1. L. Barroso, Warehouse-Scale Computers, Usenix Ann. Technical Conf., Usenix, 2007; 2. R. Bryant, Data Intensive Super Computing: Taking Google-Style Computing Beyond Web Search, Computing Community Consortium at the Federated Computing Research Conf., 2007; washington.edu/fcrc/bryant.fcrc.pdf. 3. J. Dean and S. Ghemawat, MapReduce: Simplified Data ProcessingonLargeClusters, Proc. Symp. Operating System Design and Implementation (OSDI-6), Usenix, 2004, pp P. Gill et al., YouTube Traffic Characterization: A View from the Edge, Proc. Internet Measurement Conf., ACM Press, 2007, pp
9 [3B2-3] mmi d 3/2/09 14:56 Page C. Patel and A. Shah, Cost Model for Planning, Development and Operation of a Data Center, tech. report HPL R1, Hewlett-Packard, A. Falcon, P. Faraboschi, and D. Ortega, Combining Simulation and Virtualization through Dynamic Sampling, Proc. IEEE Int l Symp. Performance Analysis of Systems and Software (ISPASS), IEEE CS Press, 2007, pp R. Bedichek, SimNow: Fast Platform Simulation Purely in Software, Proc. Hot Chips 16, IEEE CS Press, 2004; _HC16_Sess4_Pres1_bw.pdf. 8. L. Barroso, J. Dean, and U. Holzle, Web Search for a Planet: The Google Cluster Architecture, IEEE Micro, vol. 23, no. 2, 2003, pp X. Fan, W. Weber, and L. Barroso, Power Provisioning for a Warehouse-Sized Computer, Proc. 34th Int l Symp. Computer Architecture (ISCA-34), ACM Press, 2007, pp P. Ranganathan et al., Ensemble-Level Power Management for Dense Blade Servers, Proc. 33rd Int l Symp. Computer Architecture (ISCA-33), ACM Press, 2006, pp M. Ekman and P. Stenstrom, A Cost-Effective Main Organization for Future Servers, Proc. IEEE Parallel and Distributed Processing Symp. (IPDPS), IEEE Press, T. Kgil and T. Mudge, Flashcache: A NAND Flash File Cache for Low Power Web Servers, Proc. Int l Conf. Compilers, Architecture, and Synthesis for Embedded Systems (Cases), ACM Press, 2006, pp Kevin Lim is a PhD candidate in the Computer Science and Engineering Department at the University of Michigan. His research interests include data-center-level architectures, memory systems, and simulators. Lim has an MSE in computer science and engineering from the University of Michigan. Parthasarathy Ranganathan is a distinguished technologist at Hewlett-Packard Labs. His research interests include systems architecture and power efficiency. Ranganathan has a PhD in electrical and computer engineering from Rice University. He is a senior member of the IEEE and ACM. Jichuan Chang is researcher at Hewlett- Packard Labs. His research interests include computer systems architecture and memory systems. Chang has a PhD in computer science from the University of Wisconsin- Madison. He is a member of the ACM. Chandrakant Patel is a Hewlett-Packard Fellow and director of the Sustainable Information Technology Laboratory at Hewlett-Packard Labs. His research interests include creating an IT ecosystem to enable net-positive impact on the environment. Patel has an MS in mechanical engineering from San Jose State University and is a licensed professional mechanical engineer in California. He is an IEEE Fellow. Trevor Mudge is the first Bredt Family Professor of Computer Science and Engineering at the University of Michigan. His research interests include computer architecture, VLSI, and compilers. Mudge has a PhD in computer science from the University of Illinois, Urbana-Champaign. He is a member of the ACM, the Institution of Engineering and Technology, and the British Computer Society, and is a Fellow of the IEEE. Steven K. Reinhardt is a Fellow in AMD s Research and Advanced Development Labs and an adjunct associate professor at the University of Michigan. His research interests include computer system architecture. Reinhardt has a PhD in computer sciences from the University of Wisconsin-Madison. He is a member of the IEEE and ACM. Direct questions and comments about this article to Kevin Lim, Univ. of Michigan, 2260 Hayward St., CSE, Ann Arbor, MI 48109; [email protected]. For more information on this or any other computing topic, please visit our Digital Library at JANUARY/FEBRUARY
Intel RAID SSD Cache Controller RCS25ZB040
SOLUTION Brief Intel RAID SSD Cache Controller RCS25ZB040 When Faster Matters Cost-Effective Intelligent RAID with Embedded High Performance Flash Intel RAID SSD Cache Controller RCS25ZB040 When Faster
Resource Efficient Computing for Warehouse-scale Datacenters
Resource Efficient Computing for Warehouse-scale Datacenters Christos Kozyrakis Stanford University http://csl.stanford.edu/~christos DATE Conference March 21 st 2013 Computing is the Innovation Catalyst
ioscale: The Holy Grail for Hyperscale
ioscale: The Holy Grail for Hyperscale The New World of Hyperscale Hyperscale describes new cloud computing deployments where hundreds or thousands of distributed servers support millions of remote, often
Managing Data Center Power and Cooling
White PAPER Managing Data Center Power and Cooling Introduction: Crisis in Power and Cooling As server microprocessors become more powerful in accordance with Moore s Law, they also consume more power
Concepts Introduced in Chapter 6. Warehouse-Scale Computers. Important Design Factors for WSCs. Programming Models for WSCs
Concepts Introduced in Chapter 6 Warehouse-Scale Computers introduction to warehouse-scale computing programming models infrastructure and costs cloud computing A cluster is a collection of desktop computers
Virtualization of the MS Exchange Server Environment
MS Exchange Server Acceleration Maximizing Users in a Virtualized Environment with Flash-Powered Consolidation Allon Cohen, PhD OCZ Technology Group Introduction Microsoft (MS) Exchange Server is one of
Benchmarking Hadoop & HBase on Violin
Technical White Paper Report Technical Report Benchmarking Hadoop & HBase on Violin Harnessing Big Data Analytics at the Speed of Memory Version 1.0 Abstract The purpose of benchmarking is to show advantages
Flash Memory Technology in Enterprise Storage
NETAPP WHITE PAPER Flash Memory Technology in Enterprise Storage Flexible Choices to Optimize Performance Mark Woods and Amit Shah, NetApp November 2008 WP-7061-1008 EXECUTIVE SUMMARY Solid state drives
Cloud Computing Data Centers
Cloud Computing Data Centers Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Warehouse-Scale Data Center (WSC) Design. Example: a data center
MS Exchange Server Acceleration
White Paper MS Exchange Server Acceleration Using virtualization to dramatically maximize user experience for Microsoft Exchange Server Allon Cohen, PhD Scott Harlin OCZ Storage Solutions, Inc. A Toshiba
Flash Memory Arrays Enabling the Virtualized Data Center. July 2010
Flash Memory Arrays Enabling the Virtualized Data Center July 2010 2 Flash Memory Arrays Enabling the Virtualized Data Center This White Paper describes a new product category, the flash Memory Array,
Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments
Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments Applied Technology Abstract This white paper introduces EMC s latest groundbreaking technologies,
Evaluation Report: HP Blade Server and HP MSA 16GFC Storage Evaluation
Evaluation Report: HP Blade Server and HP MSA 16GFC Storage Evaluation Evaluation report prepared under contract with HP Executive Summary The computing industry is experiencing an increasing demand for
Intel DPDK Boosts Server Appliance Performance White Paper
Intel DPDK Boosts Server Appliance Performance Intel DPDK Boosts Server Appliance Performance Introduction As network speeds increase to 40G and above, both in the enterprise and data center, the bottlenecks
Support a New Class of Applications with Cisco UCS M-Series Modular Servers
Solution Brief December 2014 Highlights Support a New Class of Applications Cisco UCS M-Series Modular Servers are designed to support cloud-scale workloads In which a distributed application must run
Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database
Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database Built up on Cisco s big data common platform architecture (CPA), a
Performance Beyond PCI Express: Moving Storage to The Memory Bus A Technical Whitepaper
: Moving Storage to The Memory Bus A Technical Whitepaper By Stephen Foskett April 2014 2 Introduction In the quest to eliminate bottlenecks and improve system performance, the state of the art has continually
Cisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops
Cisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops Greater Efficiency and Performance from the Industry Leaders Citrix XenDesktop with Microsoft
Private cloud computing advances
Building robust private cloud services infrastructures By Brian Gautreau and Gong Wang Private clouds optimize utilization and management of IT resources to heighten availability. Microsoft Private Cloud
The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays
The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays Executive Summary Microsoft SQL has evolved beyond serving simple workgroups to a platform delivering sophisticated
Best Practices. Server: Power Benchmark
Best Practices Server: Power Benchmark Rising global energy costs and an increased energy consumption of 2.5 percent in 2011 is driving a real need for combating server sprawl via increased capacity and
Reduce Latency and Increase Application Performance Up to 44x with Adaptec maxcache 3.0 SSD Read and Write Caching Solutions
MAXCACHE 3. WHITEPAPER Reduce Latency and Increase Application Performance Up to 44x with Adaptec maxcache 3. SSD Read and Write Caching Solutions Executive Summary Today s data centers and cloud computing
Accelerating Server Storage Performance on Lenovo ThinkServer
Accelerating Server Storage Performance on Lenovo ThinkServer Lenovo Enterprise Product Group April 214 Copyright Lenovo 214 LENOVO PROVIDES THIS PUBLICATION AS IS WITHOUT WARRANTY OF ANY KIND, EITHER
I D C T E C H N O L O G Y S P O T L I G H T. I m p r o ve I T E f ficiency, S t o p S e r ve r S p r aw l
I D C T E C H N O L O G Y S P O T L I G H T Optimizing Blade Ser ve r s f o r V irtualization: I m p r o ve I T E f ficiency, S t o p S e r ve r S p r aw l August 2011 Adapted from IDC Blade Market Survey:
DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION
DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION A DIABLO WHITE PAPER AUGUST 2014 Ricky Trigalo Director of Business Development Virtualization, Diablo Technologies
Virtual CDNs: Maximizing Performance While Minimizing Cost
White Paper Virtual CDNs: Maximizing Performance While Minimizing Cost Prepared by Roz Roseboro Senior Analyst, Heavy Reading www.heavyreading.com on behalf of www.6wind.com www.hp.com November 2014 Introduction
Cisco Unified Computing System and EMC VNXe3300 Unified Storage System
Cisco Unified Computing System and EMC VNXe3300 Unified Storage System An Ideal Solution for SMB Server Consolidation White Paper January 2011, Revision 1.0 Contents Cisco UCS C250 M2 Extended-Memory Rack-Mount
Chapter 7: Distributed Systems: Warehouse-Scale Computing. Fall 2011 Jussi Kangasharju
Chapter 7: Distributed Systems: Warehouse-Scale Computing Fall 2011 Jussi Kangasharju Chapter Outline Warehouse-scale computing overview Workloads and software infrastructure Failures and repairs Note:
SAN Conceptual and Design Basics
TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer
OmniCube. SimpliVity OmniCube and Multi Federation ROBO Reference Architecture. White Paper. Authors: Bob Gropman
OmniCube SimpliVity OmniCube and Multi Federation ROBO Reference Architecture White Paper Authors: Bob Gropman Date: April 13, 2015 SimpliVity and OmniCube are trademarks of SimpliVity Corporation. All
EMC Unified Storage for Microsoft SQL Server 2008
EMC Unified Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON and EMC FAST Cache Reference Copyright 2010 EMC Corporation. All rights reserved. Published October, 2010 EMC believes the information
Benchmarking Cassandra on Violin
Technical White Paper Report Technical Report Benchmarking Cassandra on Violin Accelerating Cassandra Performance and Reducing Read Latency With Violin Memory Flash-based Storage Arrays Version 1.0 Abstract
Datacenter Efficiency
EXECUTIVE STRATEGY BRIEF Operating highly-efficient datacenters is imperative as more consumers and companies move to a cloud computing environment. With high energy costs and pressure to reduce carbon
System Models for Distributed and Cloud Computing
System Models for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Classification of Distributed Computing Systems
Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software
WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications
Azul Compute Appliances
W H I T E P A P E R Azul Compute Appliances Ultra-high Capacity Building Blocks for Scalable Compute Pools WP_ACA0209 2009 Azul Systems, Inc. W H I T E P A P E R : A z u l C o m p u t e A p p l i a n c
Everything you need to know about flash storage performance
Everything you need to know about flash storage performance The unique characteristics of flash make performance validation testing immensely challenging and critically important; follow these best practices
Load Balancing and Maintaining the Qos on Cloud Partitioning For the Public Cloud
Load Balancing and Maintaining the Qos on Cloud Partitioning For the Public Cloud 1 S.Karthika, 2 T.Lavanya, 3 G.Gokila, 4 A.Arunraja 5 S.Sarumathi, 6 S.Saravanakumar, 7 A.Gokilavani 1,2,3,4 Student, Department
Unified Computing Systems
Unified Computing Systems Cisco Unified Computing Systems simplify your data center architecture; reduce the number of devices to purchase, deploy, and maintain; and improve speed and agility. Cisco Unified
Cisco Unified Data Center Solutions for MapR: Deliver Automated, High-Performance Hadoop Workloads
Solution Overview Cisco Unified Data Center Solutions for MapR: Deliver Automated, High-Performance Hadoop Workloads What You Will Learn MapR Hadoop clusters on Cisco Unified Computing System (Cisco UCS
Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage
White Paper Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage A Benchmark Report August 211 Background Objectivity/DB uses a powerful distributed processing architecture to manage
CURTAIL THE EXPENDITURE OF BIG DATA PROCESSING USING MIXED INTEGER NON-LINEAR PROGRAMMING
Journal homepage: http://www.journalijar.com INTERNATIONAL JOURNAL OF ADVANCED RESEARCH RESEARCH ARTICLE CURTAIL THE EXPENDITURE OF BIG DATA PROCESSING USING MIXED INTEGER NON-LINEAR PROGRAMMING R.Kohila
Cloud computing doesn t yet have a
The Case for Cloud Computing Robert L. Grossman University of Illinois at Chicago and Open Data Group To understand clouds and cloud computing, we must first understand the two different types of clouds.
How To Build A Cloud Computer
Introducing the Singlechip Cloud Computer Exploring the Future of Many-core Processors White Paper Intel Labs Jim Held Intel Fellow, Intel Labs Director, Tera-scale Computing Research Sean Koehl Technology
Relocating Windows Server 2003 Workloads
Relocating Windows Server 2003 Workloads An Opportunity to Optimize From Complex Change to an Opportunity to Optimize There is much you need to know before you upgrade to a new server platform, and time
Cost-effective Strategies for Building the Next-generation Data Center
White Paper Cost-effective Strategies for Building the Next-generation Data Center Custom-made servers bearing energy-efficient processors are key to today s cloud computing-inspired architectures. Tom
Easier - Faster - Better
Highest reliability, availability and serviceability ClusterStor gets you productive fast with robust professional service offerings available as part of solution delivery, including quality controlled
Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and IBM FlexSystem Enterprise Chassis
White Paper Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and IBM FlexSystem Enterprise Chassis White Paper March 2014 2014 Cisco and/or its affiliates. All rights reserved. This document
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB
BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB Planet Size Data!? Gartner s 10 key IT trends for 2012 unstructured data will grow some 80% over the course of the next
Maximizing Server Storage Performance with PCI Express and Serial Attached SCSI. Article for InfoStor November 2003 Paul Griffith Adaptec, Inc.
Filename: SAS - PCI Express Bandwidth - Infostor v5.doc Maximizing Server Storage Performance with PCI Express and Serial Attached SCSI Article for InfoStor November 2003 Paul Griffith Adaptec, Inc. Server
Cloud Networking: A Novel Network Approach for Cloud Computing Models CQ1 2009
Cloud Networking: A Novel Network Approach for Cloud Computing Models CQ1 2009 1 Arista s Cloud Networking The advent of Cloud Computing changes the approach to datacenters networks in terms of throughput
Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack
Elasticsearch on Cisco Unified Computing System: Optimizing your UCS infrastructure for Elasticsearch s analytics software stack HIGHLIGHTS Real-Time Results Elasticsearch on Cisco UCS enables a deeper
Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and Dell PowerEdge M1000e Blade Enclosure
White Paper Power Efficiency Comparison: Cisco UCS 5108 Blade Server Chassis and Dell PowerEdge M1000e Blade Enclosure White Paper March 2014 2014 Cisco and/or its affiliates. All rights reserved. This
All-Flash Arrays: Not Just for the Top Tier Anymore
All-Flash Arrays: Not Just for the Top Tier Anymore Falling prices, new technology make allflash arrays a fit for more financial, life sciences and healthcare applications EXECUTIVE SUMMARY Real-time financial
Infrastructure Matters: POWER8 vs. Xeon x86
Advisory Infrastructure Matters: POWER8 vs. Xeon x86 Executive Summary This report compares IBM s new POWER8-based scale-out Power System to Intel E5 v2 x86- based scale-out systems. A follow-on report
Maximizing Hadoop Performance and Storage Capacity with AltraHD TM
Maximizing Hadoop Performance and Storage Capacity with AltraHD TM Executive Summary The explosion of internet data, driven in large part by the growth of more and more powerful mobile devices, has created
Power efficiency and power management in HP ProLiant servers
Power efficiency and power management in HP ProLiant servers Technology brief Introduction... 2 Built-in power efficiencies in ProLiant servers... 2 Optimizing internal cooling and fan power with Sea of
Cisco Nexus 5000 Series Switches: Decrease Data Center Costs with Consolidated I/O
Cisco Nexus 5000 Series Switches: Decrease Data Center Costs with Consolidated I/O Introduction Data centers are growing at an unprecedented rate, creating challenges for enterprises. Enterprise-level
Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage. White Paper
Integration of Microsoft Hyper-V and Coraid Ethernet SAN Storage White Paper June 2011 2011 Coraid, Inc. Coraid, Inc. The trademarks, logos, and service marks (collectively "Trademarks") appearing on the
Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers
WHITE PAPER FUJITSU PRIMERGY AND PRIMEPOWER SERVERS Performance Comparison of Fujitsu PRIMERGY and PRIMEPOWER Servers CHALLENGE Replace a Fujitsu PRIMEPOWER 2500 partition with a lower cost solution that
HP ProLiant BL460c achieves #1 performance spot on Siebel CRM Release 8.0 Benchmark Industry Applications running Microsoft, Oracle
HP ProLiant BL460c achieves #1 performance spot on Siebel CRM Release 8.0 Benchmark Industry Applications running Microsoft, Oracle HP ProLiant BL685c takes #2 spot HP Leadership» The HP ProLiant BL460c
The virtualization of SAP environments to accommodate standardization and easier management is gaining momentum in data centers.
White Paper Virtualized SAP: Optimize Performance with Cisco Data Center Virtual Machine Fabric Extender and Red Hat Enterprise Linux and Kernel-Based Virtual Machine What You Will Learn The virtualization
Server Platform Optimized for Data Centers
Platform Optimized for Data Centers Franz-Josef Bathe Toshio Sugimoto Hideaki Maeda Teruhisa Taji Fujitsu began developing its industry-standard server series in the early 1990s under the name FM server
Last time. Data Center as a Computer. Today. Data Center Construction (and management)
Last time Data Center Construction (and management) Johan Tordsson Department of Computing Science 1. Common (Web) application architectures N-tier applications Load Balancers Application Servers Databases
Accelerating and Simplifying Apache
Accelerating and Simplifying Apache Hadoop with Panasas ActiveStor White paper NOvember 2012 1.888.PANASAS www.panasas.com Executive Overview The technology requirements for big data vary significantly
Datacenters and Cloud Computing. Jia Rao Assistant Professor in CS http://cs.uccs.edu/~jrao/cs5540/spring2014/index.html
Datacenters and Cloud Computing Jia Rao Assistant Professor in CS http://cs.uccs.edu/~jrao/cs5540/spring2014/index.html What is Cloud Computing? A model for enabling ubiquitous, convenient, ondemand network
Certification: HP ATA Servers & Storage
HP ExpertONE Competency Model Certification: HP ATA Servers & Storage Overview Achieving an HP certification provides relevant skills that can lead to a fulfilling career in Information Technology. HP
Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS
Essentials Guide CONSIDERATIONS FOR SELECTING ALL-FLASH STORAGE ARRAYS M ost storage vendors now offer all-flash storage arrays, and many modern organizations recognize the need for these highperformance
WHITE PAPER. www.fusionstorm.com. Building Blocks of the Modern Data Center
WHITE PAPER: Easing the Way to the Cloud: 1 WHITE PAPER Building Blocks of the Modern Data Center How Integrated Infrastructure Solutions Help to Accelerate Application Deployments and Simplify Management
VDI Without Compromise with SimpliVity OmniStack and Citrix XenDesktop
VDI Without Compromise with SimpliVity OmniStack and Citrix XenDesktop Page 1 of 11 Introduction Virtual Desktop Infrastructure (VDI) provides customers with a more consistent end-user experience and excellent
Cloud Storage. Parallels. Performance Benchmark Results. White Paper. www.parallels.com
Parallels Cloud Storage White Paper Performance Benchmark Results www.parallels.com Table of Contents Executive Summary... 3 Architecture Overview... 3 Key Features... 4 No Special Hardware Requirements...
How to Cut VDI Storage Costs and Optimize Performance with VERDE Storage Optimizer Cache I/O Technology
How to Cut VDI Storage Costs and Optimize Performance with VERDE Storage Optimizer Cache I/O Technology A Virtual Bridges, Inc. Briefing Paper by Leo Reiter, Co-Founder and CTO, Virtual Bridges August
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW
HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW 757 Maleta Lane, Suite 201 Castle Rock, CO 80108 Brett Weninger, Managing Director [email protected] Dave Smelker, Managing Principal [email protected]
Enabling Technologies for Distributed and Cloud Computing
Enabling Technologies for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Multi-core CPUs and Multithreading
Overview of Next-Generation Green Data Center
Overview of Next-Generation Green Data Center Kouichi Kumon The dramatic growth in the demand for data centers is being accompanied by increases in energy consumption and operating costs. To deal with
The safer, easier way to help you pass any IT exams. Exam : 000-115. Storage Sales V2. Title : Version : Demo 1 / 5
Exam : 000-115 Title : Storage Sales V2 Version : Demo 1 / 5 1.The IBM TS7680 ProtecTIER Deduplication Gateway for System z solution is designed to provide all of the following EXCEPT: A. ESCON attach
Scaling from Datacenter to Client
Scaling from Datacenter to Client KeunSoo Jo Sr. Manager Memory Product Planning Samsung Semiconductor Audio-Visual Sponsor Outline SSD Market Overview & Trends - Enterprise What brought us to NVMe Technology
The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage
The Shortcut Guide to Balancing Storage Costs and Performance with Hybrid Storage sponsored by Dan Sullivan Chapter 1: Advantages of Hybrid Storage... 1 Overview of Flash Deployment in Hybrid Storage Systems...
Energy Constrained Resource Scheduling for Cloud Environment
Energy Constrained Resource Scheduling for Cloud Environment 1 R.Selvi, 2 S.Russia, 3 V.K.Anitha 1 2 nd Year M.E.(Software Engineering), 2 Assistant Professor Department of IT KSR Institute for Engineering
Optimize VDI with Server-Side Storage Acceleration
WHITE PAPER Optimize VDI with Server-Side Storage Acceleration Eliminate Storage Bottlenecks for Fast, Reliable Virtual Desktop Performance 1 Virtual Desktop Infrastructures (VDI) give users easy access
Chapter 13 Selected Storage Systems and Interface
Chapter 13 Selected Storage Systems and Interface Chapter 13 Objectives Appreciate the role of enterprise storage as a distinct architectural entity. Expand upon basic I/O concepts to include storage protocols.
Business white paper. HP Process Automation. Version 7.0. Server performance
Business white paper HP Process Automation Version 7.0 Server performance Table of contents 3 Summary of results 4 Benchmark profile 5 Benchmark environmant 6 Performance metrics 6 Process throughput 6
Cisco Unified Computing System: Meet the Challenges of Virtualization with Microsoft Hyper-V
White Paper Cisco Unified Computing System: Meet the Challenges of Virtualization with Microsoft Hyper-V What You Will Learn The modern virtualized data center is today s new IT service delivery foundation,
Oracle BI EE Implementation on Netezza. Prepared by SureShot Strategies, Inc.
Oracle BI EE Implementation on Netezza Prepared by SureShot Strategies, Inc. The goal of this paper is to give an insight to Netezza architecture and implementation experience to strategize Oracle BI EE
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM Sneha D.Borkar 1, Prof.Chaitali S.Surtakar 2 Student of B.E., Information Technology, J.D.I.E.T, [email protected] Assistant Professor, Information
