1 Deploying Microsoft Exchange Server 2007 mailbox roles on VMware Infrastructure 3 using HP ProLiant servers and HP StorageWorks Executive summary...2 Target audience...2 Introduction...2 Disclaimer...3 Testing overview...3 Lab configuration...4 Overview on virtual machine benchmarking...7 Phase one: I/O stress testing using Jetstress...7 Phase two: mailbox simulation using LoadGen building block test building block test building block test...15 Summary...17 Appendix A...19 Appendix B vendor support...20 VMware support statements...20 For more information...21
2 Executive summary Historically, some IT groups have been reluctant to virtualize applications such as Microsoft s Exchange Server 2003 (Exchange 2003). The performance demands Exchange 2003 placed on the server s hardware left little processor headroom for additional server consolidation. A large Exchange mailbox server could fully utilize a four-processor server. However, with the introduction of dual-core and quad-core processors, server hardware performance can now scale significantly higher. Furthermore, architectural changes in Microsoft Exchange Server 2007 have provided performance improvements that also reduce the performance cost of virtualization. For this white paper, a series of mailbox building blocks were tested using Exchange Server 2007 Enterprise Edition (Exchange 2007). The building blocks were used to measure performance of a virtualized mailbox server role using VMware Infrastructure 3 (VI3) 1. Performance was measured for a mailbox server configured as guest virtual machines (VMs) with 500, 1000 and 2000 mailboxes. Test results indicate that the performance of each of the building block configurations was acceptable in terms of CPU and I/O performance. Virtualization of the mailbox server role left a majority of the system processing power available for other work in even the largest of the Exchange Server 2007 building block configurations. Target audience The information contained in this white paper is intended for solutions architects, engineers, and project managers involved in the planning and virtualization of Microsoft Exchange Server You should be familiar with Exchange Server 2007 terminology and best practices as well as deployment and support of VMware Infrastructure 3 solutions. Introduction Using the building block approach to Exchange analysis serves two purposes. First, multiple measurements of various Exchange loads and CPU configurations provide trending data that assist in the understanding of the load characteristics. Second, CPU usage of heavily loaded, differently-sized building blocks suggests configurations when scaling out an Exchange deployment to multiple VMs on a single machine. With accurate VM sizing data, smart decisions in the tradeoff between highly-flexible, easily-migrated small VMs and minimal-maintenance large VMs can be made. For more information refer to the VMware ESX Server: A comprehensive guide on how ESX virtualizes HP ProLiant Servers white paper, located at To establish a baseline, the building blocks were initially tested using the physical server, and then repeated as a virtualized server. The Exchange Server 2007 building block testing used Microsoft s recommended heavy mailbox profile guidelines for system memory per user for database cache and for the number of mailbox users per CPU core necessary to support each of the building block configurations. Testing was performed using the HP ProLiant DL580 G4 with a total of eight CPU cores and 32GB of system memory. For more information on the ProLiant DL580 G4 server, go to html. For even more performance headroom, the recently announced HP ProLiant DL580 G5 server offers quad-core processors, which doubles the number of CPU cores to 16 and support for up to 128GB of system memory. For more information on the ProLiant DL580 G5 server, go to
3 Testing was conducted in two phases; the first phase was to measure the storage I/O performance. This was done using the Microsoft Exchange Server Jetstress tool (Jetstress). Testing was performed on the physical server and then repeated using the guest VM, to compare I/O storage performance between the two environments. The second phase of testing used the Microsoft Exchange Load Generator tool (LoadGen) to simulate messaging client traffic and was used to measure performance of the fully configured Exchange 2007 mailbox server, both as a physical server and as guest VM. Performance was measured in two key areas, CPU utilization and storage I/O performance, which are typically the most demanded resources when virtualizing an Exchange server. To provide an accurate measurement of virtual machine workloads, CPU performance metrics were collected using esxtop 2. esxtop is a VMware performance data collection tool, used to measure server performance metrics. Test results were then validated using the LoadGen pass/fail criteria to validate server performance metrics. Appendix A lists the performance counters pass/fail criteria and expected messages sent and received per user per day. Disclaimer Microsoft does not support its software running in conjunction with non-microsoft hardware virtualization software, except for Microsoft customers who have a Premier-level support agreement. For Premier-level customers, Microsoft will use commercially reasonable efforts to investigate potential issues with Microsoft software running in conjunction with non- Microsoft hardware virtualization software. For Microsoft customers who do not have a Premier-level support agreement, Microsoft will require the issue to be reproduced independently from the non-microsoft hardware virtualization software. Please refer to Microsoft s Support statement for more details For addition support details refer to Appendix B vendor support. Testing overview The test matrix was broken down into two discrete phases: 1. Storage I/O stress testing using Jetstress 2. Mailbox simulation using LoadGen Initial testing was performed to validate storage performance, for both the physical server and guest VM. Jetstress 3 is specifically designed to exercise the storage subsystem. Jetstress simulates the I/O patterns used by the Exchange Jet database engine. As a best practice, Jetstress should be used to perform I/O performance testing to validate the storage configuration before deploying any Exchange 2007 mailbox server role. After the initial storage configuration was validated using Jetstress, the Exchange 2007 topology was built and tested using LoadGen 4. Unlike Jetstress, LoadGen requires the complete Exchange topology, including the Active Directory domain, the client access server (CAS), the hub transport server (HUB) and mailbox server roles to be installed. LoadGen is used to simulate some of the key functions of the Microsoft Office Outlook 2007 client to exercise the mailbox server. Testing of the mailbox building block configurations was initially performed on the physical server (non-virtual) and then repeated as a virtual machine. To assist direct comparison, both the physical and virtual machine used the identical hardware configuration including the number of processors, memory size and storage subsystem
4 Lab configuration ProLiant DL580 G4 (server under test): Four Dual-Core 64-bit Intel Xeon 7130M Processor (3.20 GHz, 150 Watts, 800MHz FSB) 32GB PC Registered Memory (DDR2-400) with Advanced ECC functionality HP StorageWorks FC2242SR Dual Channel 4 Gb PCI-e HBA Two 72GB 10k Small form factor serial attached SCSI (SAS) drives for host operating system (OS) Two 72GB 10k SAS drives for guest OS RAID1 disk arrays for host OS disk and guest OS disk HP Smart Array P400 SAS controller with battery backed write cache (BBWC) BBWC set to 50/50 read/write Integrated NC371i Multifunction Gigabit Network Adapter, 1 Gb network Intel Virtualization Technology (Intel VT) enabled in BIOS Hyper-Threading enabled HP StorageWorks 8000 Enterprise Virtual Array (EVA8000) storage array: GB 15K fibre channel disks Two disk groups were created: one for the mailbox databases and a second for the transaction logs Local network: 1 Gb single switched network (no V-LAN) Software configuration (physical server and guest OS): Microsoft Windows Server 2003 Enterprise x64 Edition SP1 Multiprocessor hardware abstraction layer (HAL) (used for all test configurations) Hotfixes KB898060, KB and KB Exchange Server 2007 Enterprise Edition build Emulex HBA driver (physical server only) Network driver (physical server only) VMware tools (for VM s guest OS only) Video hardware acceleration (for guest OS only) LSI Logic SCSI driver (for guest OS only) Tuning settings: Boot.ini (physical server only) /NumProc will be used to define the number of processors /MaxMem will be used to define the amount of memory to be used VM settings (guest VM OS only) Number of processors (SMP) defined per VM Memory defined per VM HBA settings (physical server only) Queue depth set to 32 (default setting) Queue target set to per LUN (default setting) 4
5 Network: 1 Gb full duplex Multipath I/O (MPIO): Shortest queue service time (default setting) Windows disk administration: Basic disk partitions Default allocation size (OS disks and Log) Sector aligned database logical unit numbers (LUN) using 64k allocation units and 64 as offset VMware ESX Server (host installation): Partitions options: Installed host on SAS 72GB disk (RAID1) Used recommended disk selection Used recommended partition settings Used recommended boot specification, install on master boot record (MBR) Exchange database and logs disks were mapped as RAW devices Network: NetXtreme II 5706 Default network for VM Single v-network Support servers: Active Directory server (physical) CAS and HUB transport server (physical) LoadGen configuration: Office Outlook 2007 online mode using the heavy profile Build (web release) 5
6 Figure 1. Lab layout 6
7 Overview on virtual machine benchmarking The performance results that follow demonstrate several artifacts of virtual system performance that may be unfamiliar to those used to physical system benchmarking. All of these have to do with the manner in which CPU cycles are used and analyzed and are worth description. ESX Server executes code running on a virtual machine s virtual CPU (vcpu) in a linear manner. That is, code running on a single vcpu is not ever concurrently run on multiple physical CPUs. However, it is the prerogative of the hypervisor s scheduler to move that code between different cores or sockets for its own goals. The result of in-host CPU migrations is a distribution of vcpu work across multiple physical CPUs (pcpus), which is done to increase total system efficiency. Virtual device drivers running on the host can and are often implemented to use multiple threads. In addition, other host activities (such as the web interface, console OS time keeping, etc.) utilize non-zero CPU resources that are scheduled across the available pcpus. In short, scheduler distribution of single-threaded vcpu work, multi-threaded driver execution, and single- or multi-threaded non-vm processes can produce a distribution of work across many physical processors. In comparing virtual platform performance to the non-virtual (native) counterpart, we have chosen to include all server functions in the cost of virtualization. While this is certain to include non-essential process processor usage, it is also guaranteed to be a pessimistic bound on the maximum usage on the virtual system. CPU usage on the host will then be reported as a percentage of all available pcpus. In the native measurements, the operating system was restricted to match the number of vcpus in the VM, which means that normalization of pcpu utilization is required. One manner for doing this is to report Megacycles, which calculate work per available resources. In this manner, utilization across systems with a differing number of virtual and physical CPUs can be determined. A final word on CPU utilization: in this experiment, the Hyper-Threading feature was available on the processors. Hyper-Threading provides the appearance of two cores for each single core on the system. These two presented cores (called logical cores) provide an incremental benefit over the single real core (called a physical core). While the value of Hyper-Threading can be debated, the majority of time two logical cores provide more processing power than one physical core but less than two physical cores. Because of this non-equal processing power, CPU usage across multiple logical cores on a Hyper- Threading system is non-linear. Use of 100% of one physical core indeed represents completion of work twice as fast as 50% of one physical core. But use of 100% of two logical cores may only translate into use of as little as 110% of a single physical core. Ultimately, the presence of non-critical hypervisor activities, Hyper-Threading in the CPU, and unequal CPU counts in the physical and virtual setups makes direct comparison difficult. In short, this paper is not meant to serve as an analysis of the overhead due to virtualization but rather as a guide to the amount of utilization of this ProLiant DL580 G4 server by various building blocks. Phase one: I/O stress testing using Jetstress Phase one testing used Jetstress to validate the storage subsystem performance. The tests were initially run on the physical server and then repeated on the VM. The results were then compared to measure differences in I/O performance between the physical server and the virtual machine. Setup of the physical server Jetstress test: Boot.ini was modified to /NumProc=4 and /MaxMem=12288 Jetstress was configured with four storage groups, each with one database 7
8 The test ran with 64 threads per database (defined by auto tuning) Setup of VM Jetstress test: Memory: 12,288MB CPU: 4 SCSI Controller: LSI Logic Hard disk 1: VMware Virtual Machine Disk Format (VMDK) (OS) Hard disk 2: Raw device mapping (RDM) (database (DB)) Hard disk 3: RDM (Log) Hard disk 4: RDM (backup) Jetstress was configured with four storage groups, each with one database The test ran with 60 threads per database (defined by auto tuning) EVA8000 configuration: Database disk group 40 disks using VRAID1 Logs disk group 8 disks using VRAID1 Jetstress was configured with four storage groups (SGs) using the following system parameters: Log buffers 9000 Minimum database cache Maximum database cache 128.0MB MB Insert operations 25% Delete operations 10% Replace operations 50% Read operations 15% Lazy commits 80% The test was configured to run the two-hour performance test (random I/O) and the thread workload was defined by auto tune in both physical and VM testing. Measured areas: Achieved IOPS (I/O operations per second) on the random I/O test Disk read and write latencies on the random I/O test Data will be evaluated using the following criteria: Testing will measure IOPS for random workload Pass/Fail criteria will be based on the Jetstress test constraints Maintain read and write latency 20 ms or less 8
9 The Jetstress results in Table 1 indicate that the VM was capable of providing storage performance on parity with the similarly configured physical server. Table 1. Jetstress I/O test comparison Test Number of Disks IOPS Observed CPUs Memory Avg. Disk sec/ read Avg. Disk sec/ write Number of threads Number of SGs/ DBs IOPS per disk CPU utilization Physical GB % VM GB % As stated above, direct comparisons of CPU usage are not possible given the data presented in Table 1. However, given that on half as many Hyper-Threaded processors, the physical setup was using a little more than double the virtual setup s processors one can conclude that the CPU overhead due to virtualization is fairly small in this experiment. Table 2. esxtop Physical CPU\Processor Time Counter Average Physical CPU\0\%Processor Time 15 Physical CPU\1\%Processor Time Physical CPU\2\%Processor Time Physical CPU\3\%Processor Time Physical CPU\4\%Processor Time 13 Physical CPU\5\%Processor Time 11 Physical CPU\6\%Processor Time 11 Physical CPU\7\%Processor Time 12 Avg. % Processor time Total 7.79 Phase two: mailbox simulation using LoadGen Phase two testing consisted of a series of load generation tests to measure performance of the mailbox server role in a fully configured Exchange 2007 environment. In a manner similar to the process used in phase one, comparison testing was performed initially on the physical server and then in the VM. Testing used the Office Outlook 2007 online mode with the heavy profile to simulate a mailbox user profile. Phase two testing required a complete Exchange 2007 environment, including an Active Directory server, CAS, and HUB in addition to the mailbox server. To provide a control for the test process, only the mailbox server role was virtualized; each of the other server roles were deployed as physical servers. 9
10 OS setup (physical and guest): Windows Server 2003 Enterprise x64 Edition SP1 Multiprocessor HAL Hotfixes KB898060, KB and KB Exchange Server 2007 setup: Build Enterprise Edition Dedicated HUB and CAS server Dedicated Active Directory domain controller Mailbox physical server configured per the build block Mailbox server in VM configured per the build block One storage group/database per 500 users Content indexing off LoadGen setup: Build Test duration 12 hours test run, four-hour ramp up and eight-hour steady state Office Outlook 2007 online mode using the heavy action profile EVA8000 storage configuration: EVA storage was configured with two disk groups Database disk group was defined for each building block (as noted in each test case) Log disk group remained fixed with eight disks (minimum disk group size) for all build blocks VDisks for databases and Logs were configured using VRAID1 (disk mirroring) Database LUNs: Database LUN was sector aligned with Microsoft s Diskpar tool using a 64 KB offset Database LUN was formatted with 64 KB allocation units The three building blocks tested were sized at 500, 1000 and 2000 user mailboxes. Testing followed Microsoft s CPU core and memory guidelines for the heavy mailbox user, recommending one CPU core per 500 users and 5MB of system memory per mailbox plus an additional 2GB of memory for the operating system and the Exchange 2007 application. Measured areas: CPU and memory metrics in the physical server test using Perfmon CPU and memory metrics with in VM using esxtop Exchange performance counters including the MSExchangeIS Mailbox counter Messages Queued for Submission, and the MSExchangeIS counters RPC Average Latency and RPC Requests. Disk performance counters, including the Disk Transfer/Sec, the calculated IOPS per users, the Avg. Disk sec/read and Avg. Disk sec/write Data will be evaluated using the following criteria: CPU performance comparison on the given build blocks Disk performance to compare storage I/O Pass/Fail criteria will be based on the LoadGen metrics (see Appendix A) 10
11 500 building block test For the 500 user building block comparison testing, the guidelines recommend a single CPU core and 2GB x 5MB = 4.5GB of system memory. Physical server configuration: Boot.ini file was modified using Numproc=1 and MaxMem=4608 Virtual machine configuration: Memory: 4608MB CPU: 1 SCSI controller: LSI Logic Hard disk 1: VMware ESX Server File System (VMFS) (OS) Hard disk 2: RDM (DB) Hard disk 3: RDM (Log) Hard disk 4: RDM (backup) Exchange: One storage group with one database Initialized with 500 mailboxes using the heavy profile EVA8000 storage configuration: Database disk group was configured with 12 disks using VRAID1 Log disk group was configured with 8 disks using VRAID1 The results of the 500 user physical and virtual tests listed in Table 3 indicate acceptable disk performance and latencies and also success in terms of pass/fail criteria and expected messages sent and received per user per day (see Appendix A). Performance metrics for storage show low disk latencies below 10 ms for sec/read and less than 5 ms for sec/write. For Exchange, the three key counters remain well below the required thresholds. Table user LoadGen test Counter Physical Virtual Disk Transfers/sec IOPS/user Avg. Disk sec/read Avg. Disk sec/write Disk Read Bytes/sec 2,228,655 1,742,788 Disk Write Bytes/sec 409, ,798 MSExchangeIS Mailbox\messages queued for submission
12 Counter Physical Virtual MSExchangeIS counter\rpc Average Latency 10 ms 9 ms MSExchangeIS counter\rpc Requests 1 0 As can be seen above, there is a difference in the IO rates between the physical and virtual configurations with the virtual configuration placing lower demands on the IO subsystem. While the total work done by the Exchange server is held within acceptable boundaries, the Exchange server balances resources allocated to various components and can cause the workload to have different IO and CPU footprints. Some of the higher CPU utilization may be attributed to more aggressive cache management and is being investigated further. For the 500 user test, for the physical server, the CPU averaged 37% CPU utilization using a single processor, while the VI3 server CPU utilization is only 4.04%. However, for the virtual test, the CPU utilization is averaged across all eight processors in the server, not just the single CPU assigned to the VM. In Table 4 the individual processor times for each CPU on the VI3 server are listed along with the processor time average across all eight CPUs. Table 4. esxtop CPU processor time for guest VM with 500 users Counter Average Physical CPU\0\%Processor Time 2 Physical CPU\1\%Processor Time Physical CPU\2\%Processor Time Physical CPU\3\%Processor Time 0.05 Physical CPU\4\%Processor Time 0.05 Physical CPU\5\%Processor Time Physical CPU\6\%Processor Time 28 Physical CPU\7\%Processor Time 2 Avg. % Processor time Total 4.04 Since the processor utilization between the two test environments is not derived by the same means, it would be inaccurate to compare the 37% CPU utilization on physical server against the 4.04% measured (average across the eight CPUs) on the VI3 server. A better method to compare the CPU consumption between different test environments is to calculate the megacycles per user. Typically, the megacycles per user calculation is used to compare various client workloads and their respective CPU consumption. However, it can also be used to compare identical workloads against different processors configurations, such as dual-core and quad-core, or in this case between physical and virtual processors. 12
13 The formula of megacycles per user is: ((CPU utilization * CPU speed (MHz) * number of CPUs) / number of users under test) So, for the physical server test, the calculation is: ((0.37 * 3200 *1) / 500) = 2.36 megacycles per user For the virtual server, the calculation is: ((0.04 * 3200 * 8) / 500) = 2.04 megacycles per user The test resulted in 15% delta in megacycles between the physical server and virtual machine indicating relatively low virtualization overhead with 500 mailbox user building block configuration. As can be seen from the utilization table, utilization is confined to only one processor in the physical test. In the virtual test, you will see most of the activity on the single processor that is assigned to the virtual machine; however, you will also see some activity on all processors available to the ESX server. The additional processor activity is likely the result of systemwide activity including driver calls, idle CPU polling, and interrupts. The important point here is that a 500 user mailbox server running in a virtual machine is using only a fraction of the available processor resources, is well within performance guidelines, and offers opportunity to run additional virtual machine workloads building block test For the 1000 user building block comparison, the testing guidelines recommend two CPU cores and 2GB x 5MB = 7GB of system memory. Physical server configuration: Boot.ini was modified to /NumProc=2 and /MaxMem=7168 Virtual machine configuration: Memory: 7168MB CPU: 2 SCSI controller: LSI Logic Hard disk 1: VMFS (OS) Hard disk 2: RDM (DB) Hard disk 3: RDM (Log) Hard disk 4: RDM (backup) Exchange: Two storage groups, each with one database Initialized with 1000 mailboxes using the heavy profile 500 mailboxes per database EVA8000 storage configuration: Database disk group was configured with 22 disks using VRAID1 Log disk group was configured with 8 disks using VRAID1 The results of the 1000 user physical and virtual tests listed in Table 5 indicate acceptable disk performance and latencies and also success in terms of pass/fail criteria and message sent and received per user per day (see Appendix A). Performance metrics for storage 13
14 subsystems show low disk latencies below 10 ms sec/read and below 5 ms sec/write. For Exchange, the three key counters remain well below the require thresholds. Table user LoadGen test Counter Physical Virtual Database Cache Size (MB) 5,127 5,657 Disk Transfers/sec IOPS/user Avg. Disk sec/read Avg. Disk sec/write Disk Read Bytes/sec 4,272, 573 4,245,423 Disk Write Bytes/sec 824, ,161 MSExchangeIS Mailbox\messages queued for submission MSExchangeIS counter\rpc Average Latency ms 14 ms MSExchangeIS counter\rpc Requests 1 1 For the 1000 user test, the physical server consumed 27% of the CPU resources across the two processors, while the VI3 server consumed 10.67% across eight processors. Again the disk transfers dropped upon virtualizing the workload likely due to Exchange s tuning of the environment and use of a larger application cache. This is currently being investigated. Table 6 lists the individual CPU processor times for each of the eight CPUs on the VI3 server. Table 6. esxtop CPU processor time for guest VM with 1000 users Counter Average Physical CPU\0\%Processor Time 4 Physical CPU\1\%Processor Time Physical CPU\2\%Processor Time Physical CPU\3\%Processor Time Physical CPU\4\%Processor Time 26 Physical CPU\5\%Processor Time 4 14
15 Counter Average Physical CPU\6\%Processor Time 24 Physical CPU\7\%Processor Time 27 Avg. % Processor time across the active CPUs Using the megacycle comparison, the physical server resulted in 1.73 megacycles per user, compared with the 2.56 megacycles per user for the VM test. Nonetheless, CPU resources on ProLiant DL580 VI3 server are still considered very low at only 10.67%. In this case, the physical mailbox server is utilizing both processors equally at 23%. The ESX Server, however, has spread the majority of the workload across 3 processors in the system, with slight utilization on the remaining 5 processors. Because physical processors were not pinned exclusively to the virtual machine, ESX takes the liberty of migrating workloads to any available processor in the system to increase performance and avoid wait states. In summary, the 1000 user virtual machine utilizes a relatively small amount of available processor resources on the physical host, is well within testing guidelines, and offers opportunity for consolidation of other workloads building block test For the 2000 user building block comparison, the testing guidelines recommend four CPU cores and 2GB x 5MB = 12GB of system memory. Physical server configuration: Boot.ini was modified to /NumProc=4 and /MaxMem=12288 Virtual machine configuration: Memory: 12288MB CPU: 4 SCSI controller: LSI Logic Hard disk 1: VMFS (OS) Hard disk 2: RDM (DB) Hard disk 3: RDM (Log) Hard disk 4: RDM (backup) Exchange: Four storage groups, each with one database Initialized with 2000 mailboxes using the heavy profile 500 mailboxes per database EVA8000 storage configuration: Database disk group was configured with 40 disks using VRAID1 Log disk group was configured with 8 disks using VRAID1 The results of the 2000 user physical and virtual tests listed in Table 7 indicate acceptable disk performance and latencies and also success in terms of pass/fail criteria (see Appendix A). Performance metrics for storage show low disk latencies below 10 ms sec/read and 15
16 below 5 ms sec/write. For Exchange, the three key counters remain well below the require thresholds. Table user LoadGen test Counter Physical Virtual Database Cache Size (MB) 9,583 10,257 Disk Transfers/sec IOPS / user Avg. Disk sec/read Avg. Disk sec/write Disk Read Bytes/sec 9,715,212 8,290,040 Disk Write Bytes/sec 1,533,689 1,584,557 MSExchangeIS Mailbox\messages queued for submission MSExchangeIS counter\rpc Average Latency ms 15 ms MSExchangeIS counter\rpc Requests 1 1 Again the application tuned the environment and provided a larger cache for the virtual platform. This has some impact on the workload profile and is being investigated for later publication. For the 2000 user test the physical server consumed 28% of the CPU resources across the four processors, while the VI3 server consumed 23% across eight processors. Table 8 lists the individual processor time along with the average of 23% across all eight CPUs in the VI3 server. Table 8. esxtop CPU processor time for guest VM with 2000 users Counter Average Physical CPU\0\%Processor Time 6 Physical CPU\1\%Processor Time Physical CPU\2\%Processor Time Physical CPU\3\%Processor Time 0.12 Physical CPU\4\%Processor Time 46 Physical CPU\5\%Processor Time 46 16
17 Counter Average Physical CPU\6\%Processor Time 42 Physical CPU\7\%Processor Time 45 Avg. % Processor time across the active CPUs Using the megacycle comparison, the physical server resulted in 1.79 megacycles per user, compared with the 2.94 megacycles per user for the VM test. Regardless of comparison, CPU resources on the VI3 are still considered low at only 23.21%. In this case, the physical mailbox server is utilizing four processors equally at 25%. The ESX Server, however, has spread the majority of the workload across four processors in the system, with slight utilization on the remaining four processors. In summary, the 2,000 user virtual machine utilizes only about 50% of four processors on the physical host and is well within our testing guidelines. Summary Organizations looking to consolidate their server infrastructure looked last to Exchange due in part to Exchange 2003 demands on the server's processors and I/O storage subsystems. This essentially left little room for potential server consolidation on older two and four processor servers. However, with the introduction of dual-core and quad-core processors, server hardware can now scale significantly higher. Furthermore, architectural changes in Microsoft Exchange Server 2007 have provided performance improvements that also reduce the performance cost of virtualization The building blocks tested in this document demonstrate the performance impact or added overhead of moving the Exchange 2007 mailbox role to a virtualized environment. As previously discussed, it is difficult to do a direct processor utilization comparison between the physical server and virtual server tests. Tables 9 and 10 summarize the 500, 1000 and 2000 building block test results. Due to the non-comparable nature of the virtual and physical measurements, each is represented in is own table. In all building blocks, the overall CPU demand on the server is considered quite low. The 2000 virtualized mailbox build block averaged only 23% CPU utilization on the eight core server, providing additional head room for further consolidation. Table 9. Physical workload building block test summary Type Number of users Processors Processor utilization Physical % Physical % Physical % Table 10. Virtual building block test summary Type Number of users Processors Processor utilization 17
18 Type Number of users Processors Processor utilization VM VM / % VM VM/ % VM VM/ % 18
19 Appendix A Table 10 lists key counters and their respective pass/fail thresholds, used to determine if the LoadGen test was considered successful. Table 10. LoadGen pass/fail criteria Object and Counter Processor: % Processor Time (_Total) MSExchangeIS: RPC Requests MSExchangeIS: RPC Averaged Latency Logical disk: Average Sec/Read Logical disk Average Sec/Write MSExchangeIS Mailbox: Message Queued for Submission Expected Value Average less than 90 Average less than 50 and max of 100 Average less than 50 ms and max of 100 ms Less than 50 ms at all times. Less than 50 ms at all times. Average less than 250 and max of 1000 Table 11 lists the expected message workload using LoadGen s Office Outlook 2007 online mode heavy profile Table 11. Message sent and received per user per day Counter Expected Value Error threshold Messages Submitted/User/Day: Messages Delivered/User/Day: Recipients/message ratio 18 +/-- 5% 31 +/-- 5% /-- 5% 19
20 Appendix B vendor support VMware requires that Support and Subscription services (SnS) must be included with every VMware virtualization software sale. For more information on HP s customer technical support options for VMware Virtualization software go to: VMware support statements For more information on VMware s customer technical support options for Microsoft products running within VMware virtual machines go to: 20
21 For more information For more information on planning, deploying, or managing Microsoft Exchange Server on ProLiant servers and HP storage see: For more information on planning, deploying, or managing a virtual infrastructure on ProLiant servers see: For more information on HP ProLiant servers see: For more information on HP storage solutions see: To help us improve our documents, please provide feedback at Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. The only warranties for HP products and services are set forth in the express warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional warranty. HP shall not be liable for technical or editorial errors or omissions contained herein. Intel and Xeon are trademarks of Intel Corporation in the U.S. and other countries. Microsoft and Windows are U.S. registered trademarks of Microsoft Corporation. 4AA1-5890ENW, Revision 2, October 2007
Microsoft Exchange Server 2007 and Hyper-V high availability configuration on HP ProLiant BL680c G5 server blades Executive summary... 2 Introduction... 2 Exchange 2007 Hyper-V high availability configuration...
Sizing guide for SAP and VMware ESX Server running on HP ProLiant x86-64 platforms Executive summary... 2 Server virtualization overview... 2 Solution definition...2 SAP architecture... 2 VMware... 3 Virtual
Performance brief for IBM WebSphere Application Server.0 with VMware ESX.0 on HP ProLiant DL0 G server Table of contents Executive summary... WebSphere test configuration... Server information... WebSphere
Written and Provided by Expert Reference Series of White Papers Microsoft Exchange Server 200 Performance on VMware vsphere 4 1-800-COURSES www.globalknowledge.com Introduction Increased adoption of VMware
HP recommended configuration for Microsoft Exchange Server 2010: HP LeftHand P4000 SAN Table of contents Executive summary... 2 Introduction... 2 Solution criteria... 3 Hyper-V guest machine configurations...
Storage I/O Performance on VMware vsphere 5.1 over 16 Gigabit Fibre Channel Performance Study TECHNICAL WHITE PAPER Table of Contents Introduction... 3 Executive Summary... 3 Setup... 3 Workload... 4 Results...
Design and Sizing Examples: Microsoft Exchange Solutions on VMware Page 1 of 19 Contents 1. Introduction... 3 1.1. Overview... 3 1.2. Benefits of Running Exchange Server 2007 on VMware Infrastructure 3...
MS EXCHANGE SERVER ACCELERATION IN VMWARE ENVIRONMENTS WITH SANRAD VXL Dr. Allon Cohen Eli Ben Namer firstname.lastname@example.org 1 EXECUTIVE SUMMARY SANRAD VXL provides enterprise class acceleration for virtualized
BlackBerry Enterprise Server for Microsoft Exchange Version: 5.0 Service Pack: 4 Performance Benchmarking Guide Published: 2015-01-13 SWD-20150113132750479 Contents 1 BlackBerry Enterprise Server for Microsoft
Technical white paper HP ProLiant DL380p Gen8 1000 mailbox 2GB mailbox resiliency Exchange 2010 storage solution Table of contents Overview 2 Disclaimer 2 Features of the tested solution 2 Solution description
HP SN1000E 16 Gb Fibre Channel HBA Evaluation Evaluation report prepared under contract with Emulex Executive Summary The computing industry is experiencing an increasing demand for storage performance
How to configure Failover Clustering for Hyper-V hosts on HP ProLiant c-class server blades with All-in-One SB600c storage blade Executive summary... 2 System requirements... 2 Hardware requirements...
MS Exchange Server Acceleration Maximizing Users in a Virtualized Environment with Flash-Powered Consolidation Allon Cohen, PhD OCZ Technology Group Introduction Microsoft (MS) Exchange Server is one of
Technical Paper Moving SAS Applications from a Physical to a Virtual VMware Environment Release Information Content Version: April 2015. Trademarks and Patents SAS Institute Inc., SAS Campus Drive, Cary,
VMware vsphere 6 and Oracle Database Scalability Study Scaling Monster Virtual Machines TECHNICAL WHITE PAPER Table of Contents Executive Summary... 3 Introduction... 3 Test Environment... 3 Virtual Machine
Achieving a Million I/O Operations per Second from a Single VMware vsphere 5.0 Host Performance Study TECHNICAL WHITE PAPER Table of Contents Introduction... 3 Executive Summary... 3 Software and Hardware...
White Paper MS Exchange Server Acceleration Using virtualization to dramatically maximize user experience for Microsoft Exchange Server Allon Cohen, PhD Scott Harlin OCZ Storage Solutions, Inc. A Toshiba
Virtualized SQL Server Database Performance Study TECHNICAL WHITE PAPER Table of Contents Executive Summary... 3 Introduction... 3 Factors Affecting Storage vmotion... 3 Workload... 4 Experiments and Results...
HP and Mimosa Systems A system for email archiving, recovery, and storage optimization white paper Mimosa NearPoint for Microsoft Exchange Server and HP StorageWorks 1510i Modular Smart Array Executive
HP StorageWorks MPX200 Simplified Cost-Effective Virtualization Deployment Executive Summary... 2 HP StorageWorks MPX200 Architecture... 2 Server Virtualization and SAN based Storage... 3 VMware Architecture...
EMC Solutions for Microsoft Exchange 2007 Virtualized Exchange 2007 Local Continuous Replication EMC Commercial Solutions Group Corporate Headquarters Hopkinton, MA 01748-9103 1-508-435-1000 www.emc.com
Scaling in a Hypervisor Environment Richard McDougall Chief Performance Architect VMware VMware ESX Hypervisor Architecture Guest Monitor Guest TCP/IP Monitor (BT, HW, PV) File System CPU is controlled
Monitoring Databases on VMware Ensure Optimum Performance with the Correct Metrics By Dean Richards, Manager, Sales Engineering Confio Software 4772 Walnut Street, Suite 100 Boulder, CO 80301 www.confio.com
Analysis of VDI Storage Performance During Bootstorm Introduction Virtual desktops are gaining popularity as a more cost effective and more easily serviceable solution. The most resource-dependent process
Performance Study Performance of Virtualized SQL Server Based VMware vcenter Database VMware vsphere 4.1 VMware vsphere is a sound platform on which to virtualize SQL Server databases. One overlooked database
SAP database backup and restore solutions for HP StorageWorks Enterprise Virtual Array using HP Data Protector 6.1 software Table of contents Executive summary... 2 Solution overview... 2 Objectives...
W H I T E P A P E R Performance and Scalability of Microsoft SQL Server on VMware vsphere 4 Table of Contents Introduction................................................................... 3 Highlights.....................................................................
Technical white paper HP ProLiant BL660c Gen9 and Microsoft SQL Server 2014 technical brief Scale-up your Microsoft SQL Server environment to new heights Table of contents Executive summary... 2 Introduction...
Parallels VDI Solution White Paper VDI SIZING A Competitive Comparison of VDI Solution Sizing between Parallels VDI versus VMware VDI www.parallels.com Parallels VDI Sizing. 29 Table of Contents Overview...
Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820 This white paper discusses the SQL server workload consolidation capabilities of Dell PowerEdge R820 using Virtualization.
Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card Version 1.0 April 2011 DB15-000761-00 Revision History Version and Date Version 1.0, April 2011 Initial
Microsoft Exchange Server 3 Deployment Considerations for Small and Medium Businesses A Dell PowerEdge server can provide an effective platform for Microsoft Exchange Server 3. A team of Dell engineers
EMC Solutions for Microsoft Exchange 2007 Virtualized Exchange 2007 Archiving with EMC EmailXtender/DiskXtender to EMC Centera EMC Commercial Solutions Group Corporate Headquarters Hopkinton, MA 01748-9103
Enabling VMware Enhanced VMotion Compatibility on HP ProLiant servers Technical white paper Table of contents Executive summary... 2 VMware VMotion overview... 2 VMotion CPU compatibility... 2 Enhanced
White Paper Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study 2012 Cisco and/or its affiliates. All rights reserved. This
Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays Database Solutions Engineering By Murali Krishnan.K Dell Product Group October 2009
Virtual SAN Design and Deployment Guide TECHNICAL MARKETING DOCUMENTATION VERSION 1.3 - November 2014 Copyright 2014 DataCore Software All Rights Reserved Table of Contents INTRODUCTION... 3 1.1 DataCore
HP Asset Manager Asset Manager 5.10 Sizing Guide Using the Oracle Database Server, or IBM DB2 Database Server, or Microsoft SQL Server Legal Notices... 2 Introduction... 3 Asset Manager Architecture...
HP VMware ESXi 5.0 and Updates Getting Started Guide Abstract This guide is intended to provide setup information for HP VMware ESXi. HP Part Number: 616896-002 Published: August 2011 Edition: 1 Copyright
An evaluation of blade server power efficiency for the, Dell PowerEdge M600, and IBM BladeCenter HS21 using the SPECjbb2005 Benchmark The HP Difference The ProLiant BL260c G5 is a new class of server blade
Dell PowerVault MD32xx Deployment Guide for VMware ESX4.1 Server A Dell Technical White Paper PowerVault MD32xx Storage Array www.dell.com/md32xx THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND
Server and Storage Sizing Guide for Windows 7 TECHNICAL NOTES Table of Contents About this Document.... 3 Introduction... 4 Baseline Existing Desktop Environment... 4 Estimate VDI Hardware Needed.... 5
Virtualization Technologies and Blackboard: The Future of Blackboard Software on Multi-Core Technologies Kurt Klemperer, Principal System Performance Engineer email@example.com Agenda Session Length:
WHITE PAPER Optimizing Virtual Platform Disk Performance Think Faster. Visit us at Condusiv.com Optimizing Virtual Platform Disk Performance 1 The intensified demand for IT network efficiency and lower
Power Comparison of Dell PowerEdge 2950 using Intel X5355 and E5345 Quad Core Xeon Processors By Scott Hanson and Todd Muirhead Dell Enterprise Technology Center Dell Enterprise Technology Center dell.com/techcenter
Virtualizing Microsoft SQL Server 2008 on the Hitachi Adaptable Modular Storage 2000 Family Using Microsoft Hyper-V Implementation Guide By Eduardo Freitas and Ryan Sokolowski February 2010 Summary Deploying
Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture Dell Compellent Product Specialist Team THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL
Power Regulator for ProLiant servers Technology brief, 5th edition Introduction... 2 Processor P-states... 2 Power Regulator technology... 2 HP Power Regulator modes... 2 Processor drivers... 3 HP Static
IOmark- VDI HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC- 150427- b Test Copyright 2010-2014 Evaluator Group, Inc. All rights reserved. IOmark- VDI, IOmark- VM, VDI- IOmark, and IOmark
Improved Virtualization Performance with 9th Generation Servers David J. Morse Dell, Inc. August 2006 Contents Introduction... 4 VMware ESX Server 3.0... 4 SPECjbb2005... 4 BEA JRockit... 4 Hardware/Software
Performance Study Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build 164009 Introduction With more and more mission critical networking intensive workloads being virtualized
Performance Study Microsoft Office SharePoint Server 2007 Performance on VMware vsphere 4.1 VMware vsphere 4.1 One of the key benefits of virtualization is the ability to consolidate multiple applications
Directions for VMware Ready Testing for Application Software Introduction To be awarded the VMware ready logo for your product requires a modest amount of engineering work, assuming that the pre-requisites
April 23 11 Aviation Parkway, Suite 4 Morrisville, NC 2756 919-38-28 Fax 919-38-2899 32 B Lakeside Drive Foster City, CA 9444 65-513-8 Fax 65-513-899 www.veritest.com firstname.lastname@example.org Microsoft Windows
Virtual Desktop Infrastructure (VDI) made Easy HOW-TO Preface: Desktop virtualization can increase the effectiveness of information technology (IT) teams by simplifying how they configure and deploy endpoint
8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments QLogic 8Gb Adapter Outperforms Emulex QLogic Offers Best Performance and Scalability in Hyper-V Environments Key Findings The QLogic
Best Practices for Monitoring Databases on VMware Dean Richards Senior DBA, Confio Software 1 Who Am I? 20+ Years in Oracle & SQL Server DBA and Developer Worked for Oracle Consulting Specialize in Performance
Evaluation Report: Supporting Multiple Workloads with the Lenovo S3200 Storage Array Evaluation report prepared under contract with Lenovo Executive Summary Virtualization is a key strategy to reduce the
Technical white paper HP Smart Array Controllers and basic RAID performance factors Technology brief Table of contents Abstract 2 Benefits of drive arrays 2 Factors that affect performance 2 HP Smart Array
Squeezing The Most Performance from your VMware-based SQL Server PASS Virtualization Virtual Chapter February 13, 2013 David Klee Solutions Architect (@kleegeek) About HoB Founded in 1998 Partner-Focused
White Paper Recording Server Virtualization Prepared by: Mike Sherwood, Senior Solutions Engineer Milestone Systems 23 March 2011 Table of Contents Introduction... 3 Target audience and white paper purpose...
VMware ESX Server Multiple Workload Performance on Dell PowerEdge 2850 and PowerEdge 6850 Servers Enterprise Product Group (EPG) Dell White Paper By Todd Muirhead, Dave Jaffe and Scott Stanford July 2005
Using EonStor FC-host Storage Systems in VMware Infrastructure 3 and vsphere 4 Application Note Abstract This application note explains the configure details of using Infortrend FC-host storage systems
Performance Study PVSCSI Storage Performance VMware ESX 4.0 VMware vsphere 4 offers Paravirtualized SCSI (PVSCSI), a new virtual storage adapter. This document provides a performance comparison of PVSCSI
Course ID VMW200 VMware vsphere 5.1 Advanced Administration Course Description This powerful 5-day 10hr/day class is an intensive introduction to VMware vsphere 5.0 including VMware ESX 5.0 and vcenter.
Technical white paper HP MSA 2040 Storage 750 Mailbox Resiliency Exchange 2013 Storage Solution with Microsoft Hyper-V Virtualized Microsoft Exchange 2013 server deployment Table of contents Overview...
DELL Virtual Desktop Infrastructure Study END-TO-END COMPUTING Dell Enterprise Solutions Engineering 1 THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL
Optimizing SQL Server Storage Performance with the PowerEdge R720 Choosing the best storage solution for optimal database performance Luis Acosta Solutions Performance Analysis Group Joe Noyola Advanced
White Paper Virtualization Performance on SGI UV 2000 using Red Hat Enterprise Linux 6.3 KVM September, 2013 Author Sanhita Sarkar, Director of Engineering, SGI Abstract This paper describes how to implement
TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer