SlideShare una empresa de Scribd logo
1 de 19
Descargar para leer sin conexión
Microsoft Exchange Server 2003
Performance on VMware® ESX Server 3
                           VMware ESX Server 3
Microsoft Exchange Server 2003 Performance on VMware® ESX Server 3




Microsoft Exchange Server 2003 Performance on VMware® ESX Server 3
Revision: 20070705
Item: ESX-ENG-Q207-388




You can find the most up-to-date technical documentation on our Web site at

     http://www.vmware.com/support/

The VMware Web site also provides the latest product updates.
If you have comments about this documentation, submit your feedback to:

     docfeedback@vmware.com




© 2007 VMware, Inc. All rights reserved. Protected by one or more of U.S. Patent Nos. 6,397,242, 6,496,847, 6,704,925,
6,711,672, 6,725,289, 6,735,601, 6,785,886, 6,789,156, 6,795,966, 6,880,022, 6,944,699, 6,961,806, 6,961,941, 7,069,413,
7,082,598, 7,089,377, 7,111,086, 7,111,145, 7,117,481, 7,149,843 and 7,155,558; patents pending.
VMware, the VMware “boxes” logo and design, Virtual SMP and VMotion are registered trademarks or trademarks of VMware,
Inc. in the United States and/or other jurisdictions. Dell and PowerEdge are registered trademarks of Dell, Inc. All other marks
and names mentioned herein may be trademarks of their respective companies.




VMware, Inc.                                                                             Dell, Inc.
3145 Porter Drive                                                                        www.dell.com/vmware
Palo Alto, CA 94304
www.vmware.com




ii                                                                                                                   VMware, Inc.
Contents



          Microsoft Exchange Server 2003 Performance on VMware ESX Server 3   1
               Introduction 1
               Experiment Configuration and Methodology 2
                  Test Bed Configuration 2
                  Test and Measurement Tools 5
                  Test Cases and Test Method 5
               Experimental Results and Performance Analysis   7
                  Physical vs. Virtual Performance 7
                  Scale Up and Scale Out Performance 7
                     UP or SMP Virtual Machines? 8
                     Response Time Overheads 8
               System Component Performance 10
                  Processor Performance 10
                  Memory Performance 10
                  Networking Performance 12
                  Storage Performance 12
               Conclusion 14
               Resources and References 15
                  VMware 15
                  Dell 15
                  Microsoft 15




VMware, Inc.                                                                      iii
Virtualizing Microsoft Exchange 2003 Deployments on VMware ESX Server 3




iv                                                                        VMware, Inc.
Microsoft Exchange Server 2003 Performance
on VMware ESX Server 3


Introduction
          Virtualization has become a mainstream technology, allowing enterprises to consolidate underutilized servers
          while helping to increase reliability and fault tolerance and simplify load balancing. As organizations embrace
          virtualization in the data center, many may consider virtualizing Microsoft Exchange software. This paper
          suggests how an enterprise-critical messaging application like Microsoft Exchange Server 2003 should be
          sized and deployed on VMware® ESX Server1 to obtain a satisfactory Quality of Service. Specifically, we
          examine:

               The performance implications of running Exchange Server 2003 on a virtual machine versus a physical
               system.

               The performance of Exchange Server 2003 in virtual machine configurations when “scaling-up” (adding
               more processors to a machine) and “scaling-out” (adding more machines).

          This paper discusses the performance and scalability of Exchange Server 2003 when it is deployed within
          virtual machines hosted by VMware ESX Server 3.0.1 on a Dell® PowerEdge® 6850 server with a Dell-EMC
          CX500 FC SAN. The Heavy user profile from Microsoft’s Exchange Server 2003 Load Simulator benchmarking
          tool was used to simulate the Exchange workload. Results indicated that a uniprocessor virtual machine can
          support up to 1,300 Heavy users. Our experiments also show that consolidating multiple instances of these
          uniprocessor Exchange virtual machines on a PowerEdge 6850 can cumulatively support up to 4,000 Heavy
          users while still providing acceptable performance and scaling.

          A key observation in the study is that uniprocessor virtual machines are, from a performance perspective,
          equivalent to half as many multiprocessor (two virtual processors) virtual machines. Hence we recommend
          that the Windows and Exchange licensing costs, ease of management, and corporate standards guide your
          configuration in this regard.




          1.        When considering the deployment of Microsoft Exchange Server 2003 on a virtualization platform such as
          VMware ESX Server, please refer to http://support.microsoft.com/kb/897615 for information about the Microsoft support
          policy for their applications running in non-Microsoft virtualization software.


VMware, Inc.                                                                                                                      1
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



Experiment Configuration and Methodology
          The performance and sizing studies were done at the Dell Enterprise Solutions Engineering Labs in
          collaboration with VMware engineers. The purpose of the tests was to measure, analyze, and understand the
          performance of Exchange in both the physical and virtual environments. In the following sections the test bed
          configuration used for the experiments is described in detail, and the test tools are discussed. Finally, we
          present a description of the experiments.


     Test Bed Configuration
          In our test configuration the system under test (SUT) was a Dell PowerEdge 6850 server. The PowerEdge 6850
          was configured with four 2.66 GHz dual-core Intel Xeon 7020 processors and 16GB of RAM.

          In the physical environment, the system was running Microsoft Windows Server 2003 Enterprise Edition
          (32-bit) and Microsoft Exchange Server 2003 with Service Pack 2 (SP2). In the virtualized environment, the
          system was running VMware ESX Server 3.0.1, and the virtual machines were configured with the same
          operating system and application as in the physical system.

          The client systems were Dell PowerEdge 1855 blade servers, each with two 3.6 GHz Intel Xeon processors and
          2GB of RAM. One client system was deployed for each Exchange mailbox store. The client systems were
          running Microsoft Windows Server 2003 Enterprise Edition (32-bit) and Microsoft Exchange Server 2003 Load
          Simulator (LoadSim). In addition to its function as a mail client, the first of these client systems also served as
          the active directory server, the domain controller, and the DNS server.

          Table 1 through Table 3 summarize the test setup in more detail.
          Table 1. System Under Test (SUT) Configuration
          Server                                           Dell PowerEdge 6850
          Processors                                       Four 2.66 GHz dual-core Intel Xeon 7020 (eight total cores)

          Memory                                           16GB DDR-2 400 ECC SDRAM

          Hard drives (for operating system)               Two 146GB 10,000 RPM drives in a RAID 1 array
          HBA                                              Two Emulex Fibre Channel HBAs

          Virtualization software                          VMware ESX Server 3.0.1

          Operating system (physical and virtual)          Microsoft Windows Server 2003 Enterprise Edition (32-bit)

          Application (physical and virtual)               Microsoft Exchange Server 2003 with Service Pack 2




          Table 2. Storage Configuration
          Storage enclosure         Dell EMC CX500 with one disk processor enclosure (DPE)
                                    and three disk array enclosures (DAE)

          Hard drives               Fifty-four 73GB 15,000 RPM Drives

          RAID configuration        Eight 6-drive RAID 1+0 volumes for
                                    Microsoft Exchange Information Store (IS)


                                    One 6-drive RAID 1+0 volume partitioned into eight LUNs
                                    for the transaction log




2                                                                                                                        VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3




          Table 3. Client Configuration
          Client                  Dell PowerEdge 1855 with four blade servers

          Processors              Two 3.6 GHz Intel Xeon per blade

          Memory                  2GB per blade

          Operating system        Microsoft Windows Server 2003 Enterprise Edition (32-bit)

          Benchmark application   Microsoft Exchange Server 2003 Load Simulator (LoadSim)




VMware, Inc.                                                                                                                  3
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3




                                                                    Client
                                                                    (Dell® PowerEdge® 1855
                                                                    blade servers)




                                                                    Gigabit Ethernet switch




                                                                    Server
                                                                    (Dell® PowerEdge® 6850)




                                                                    Fibre Channel switch
                                                                    (Brocade Silkworm 3900)




                                                                    Storage
                                                                    (Dell/EMC CX500)




          Figure 1. Test Configuration Diagram



4                                                                                   VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



     Test and Measurement Tools
          Microsoft LoadSim 2003, running on the client systems, was used to simulate messaging load for the physical
          and virtual Exchange deployments. LoadSim provided the tools to measure how a system would respond
          when supporting a given number of Exchange users, in this case Messaging Application Programming
          Interface (MAPI) clients simulated using the LoadSim Heavy user profile. For a LoadSim run to be valid, 95%
          of the response times experienced by these users must be less than 500ms. This ensures that the end users get
          reasonable interactive performance.

          The benchmark tries to closely model the normal daily email usage of real users in order to provide an estimate
          of the number of users a system can support. While this benchmark is widely used to measure the performance
          of Exchange platforms, as with all benchmarks the results may not match the specifics of your environment.

          We used Windows Perfmon, a performance monitoring tool for Microsoft Windows operating systems, to
          monitor performance on the physical Exchange setups. We configured Perfmon to log relevant CPU, memory,
          disk, network, and system counters as well as Exchange-specific counters. These counters were logged to a
          comma-separated value (CSV) performance log which was later imported to a spreadsheet for analysis.

          We used esxtop, a performance monitoring tool for ESX Server, to record both ESX Server and virtual
          machine related performance counters for the virtualized test cases. We configured esxtop to log processor,
          memory, disk, network, and system counters during the LoadSim runs.

          The data from Perfmon and esxtop are the primary focus of our analysis.


     Test Cases and Test Method
          We had two primary objectives in performing these experiments:

               Comparing the performance of physical Exchange installations with their virtual counterparts.

               Understanding the scalability of the virtual machines (VMs) with respect to the number of VMs and the
               Virtual CPU count.

          We therefore created the following test cases:

               Physical uniprocessor (UP physical)

               Physical symmetric multiprocessor with two processors (SMP physical)

               One VM with one virtual CPU (UP VM)

               Two UP VMs
               Four UP VMs

               One VM with two virtual CPUs (SMP VM)

               Two SMP VMs

          As shown in Figure 4, these configurations can be compared in several ways to understand the virtualization
          overhead, “scale up” potential, and “scale out” potential.


          Table 4. Test Cases
          Comparison                                 Measurement

          UP physical     vs.   UP VM                Virtualization overhead

          SMP physical    vs.   SMP VM               Virtualization overhead

          UP VM           vs.   SMP VM               Scale up potential

          One UP VM       vs.   Two UP VMs           Scale out potential

          One UP VM       vs.   Four UP VMs          Scale out potential

          One SMP VM      vs.   Two SMP VMs          Scale out potential




VMware, Inc.                                                                                                                    5
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



          Our goal for each configuration was to find the number of LoadSim Heavy users that the configuration could
          support while continuing to meet the following Quality of Service (QoS) metrics:

               The host system CPU utilization must average 80%. This matches the typical maximum load in
               production environments and enables benchmark score comparison across configurations.

               95% of the response times must be below 500ms (as required for a LoadSim run to be valid).

          Each test was performed twice in order to verify the consistency and repeatability of the results.

          To benchmark the UP and SMP physical configurations we booted the 8-way physical box with one and two
          processors, respectively, and with 4GB of memory. This was achieved by appropriately setting the parameters
          /NUMPROC and /MAXMEM in the Windows C:boot.ini system file. Being a 32-bit application that doesn't
          support Physical Address Extension (PAE), a single instance of Exchange Server 2003 is limited to using at
          most 4GB of physical memory. ESX Server was configured to use the entire 16GB of memory available on the
          system, but each VM running on ESX Server was configured to use 4GB to match the physical configuration.

          In both the physical and virtual environments the /PAE and /3GB options in the C:boot.ini system file
          were used, the later option allowing Exchange to address up to 3GB of memory.




6                                                                                                              VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



Experimental Results and Performance Analysis
          In this section we discuss the results obtained from the experiments described above. First we present the
          measurements for comparable physical and virtual systems, then we discuss the scale up and scale out
          performance as seen in virtual environments.


     Physical vs. Virtual Performance
          In order to compare the performance of physical and virtual environments, we conducted UP and SMP
          LoadSim experiments, first on the physical system, then in the ESX Server 3.0.1 virtual environment. The
          software stacks, Windows Server 2003 running Exchange Server 2003, were configured identically on the
          physical and virtual systems. Doing so enabled an “apples to apples” comparison between the physical and
          virtual setups. The results of these experiments are shown in Figure 2.


                                                          3,500

                                                                     Virtual
                                                          3,000
                                                                     Physical
                                    LoadSim Heavy Users




                                                          2,500


                                                          2,000


                                                          1,500


                                                          1,000


                                                           500


                                                             0
                                                                  One Processor            Two Processors

          Figure 2. UP and SMP Physical and Virtual Throughput Comparisons

          In the UP case we recorded scores of 1,300 and 1,700 users for the virtual and physical setups, respectively,
          which is a virtual to physical ratio of about 0.76. In the SMP case the virtual and physical environments
          managed to support 2,200 and 3,100 users, respectively, which is a virtual to physical ratio of about 0.71.


     Scale Up and Scale Out Performance
          In addition to comparing the virtual to physical performance, we also wanted to study the scale out
          performance (by increasing the number of VMs) and the scale up performance (by increasing the number of
          virtual CPUs configured for each VM). We conducted experiments using the following configurations:

          Scale out (increasing the number of VMs):

               One, two, and four uniprocessor virtual machines.

               One and two dual-processor virtual machines.

          Scale up (increasing the number of processors per VM):

               Uniprocessor and dual-processor virtual machines.




VMware, Inc.                                                                                                                              7
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



          The results of these scale up and scale out experiments are shown in Figure 3 and detailed in Table 5.



                                 4,500

                                 4,000

                                 3,500
               Number of Users



                                 3,000

                                 2,500

                                 2,000

                                 1,500

                                 1,000

                                  500

                                    0
                                         One 1-vCPU VM    One 2-vCPU VM       Two 1-vCPU VMs   Two 2-vCPU VMs   Four 1-vCPU VMs


          Figure 3. Virtual Scale Out and Scale Up Performance


          Table 5. Scale Out and Scale Up Results
          Experiment                               First Data Point                 Second Data Point            Scaling
          Uniprocessor scale out                   One UP VM (1,300 users)          Two UP VMs (2,400 users)     1.85

          Uniprocessor scale out                   One UP VM (1,300 users)          Four UP VMs (4,000 users)    3.08

          Dual processor scale out                  One SMP VM (2,200 users)        Two SMP VMs (4,000 users)    1.82
          Processor scale up                        One UP VM (1,300 users)         One SMP VM (2,200 users)     1.69


          UP or SMP Virtual Machines?
          A fixed number of virtual CPUs (vCPUs) can be deployed in a number of ways. For instance, two VMs – each
          with one vCPU – might be compared to a single VM with two vCPUs. Each scenario has the same number of
          vCPUs, but may experience different performance depending on application characteristics. Two VMs
          configured with one vCPU each generated a score of 2,400 LoadSim Heavy users, while one VM with two
          vCPUs scored 2,200 LoadSim Heavy users. This might suggest that configuring multiple UP VMs improves
          performance compared to fewer SMP VMs. However the scenario with four uniprocessor VMs and the
          scenario with two dual-processor VMs each generated a score of 4,000 LoadSim Heavy users. So both the UP
          and SMP configurations performed approximately equally in our experiments.

          It should be noted that considerations such as licensing costs for the different configurations may result in
          different price-performance recommendations. In addition to the software licensing costs, the ease of
          management (with fewer VMs) or corporate standards (standardizing on a UP or SMP image) might dictate
          the choice of UP VMs or half as many SMP VMs.

          Response Time Overheads
          The CPU, memory, networking, and storage virtualization overheads induce higher response times for the
          Exchange benchmark running within the virtual machines compared to the physical setup. A fair comparison




8                                                                                                                            VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



          using LoadSim with the same number of Exchange users both physically and within the VMs results in the
          95th percentile response times shown in Table 6.
          Table 6. Response Time Overheads
                                         Virtual       Physical

          One processor (1,300 users)    365ms         292ms

          Two processors (2,200 users)   440ms         260ms


          The data shows an increase in the response time for the virtual setup over the physical one. But it bears
          repeating that even with an increased latency of around 100-200ms, the virtual environments still met the
          500ms QoS requirement.




VMware, Inc.                                                                                                                  9
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



System Component Performance
          The previous section discussed the benchmark throughput across the various virtual and physical
          configurations of interest. This section will analyze the impact of those throughputs on the system components
          including the processors, memory, network, and storage.


     Processor Performance
          As previously mentioned, all of the physical and virtual configurations were pushed to 80% average host
          processor utilization and the LoadSim throughputs at that CPU limit were measured. The difference in the
          virtual and physical scores can be attributed to the processor virtualization overheads combined with
          memory, storage, and networking overheads.

          To get a clearer understanding of just the processor virtualization overhead we re-ran the physical UP and
          physical SMP experiments with the same number of users as the maximum we had achieved with the UP VM
          and SMP VM virtual environments (1,300 and 2,200 users, respectively). The host processor utilizations for
          both the virtual and physical configurations are shown in Figure 4.


                                                          100%
                                                                 Physical
                                                          90%
                                                                 Virtual
                                                          80%

                                                          70%
                                        CPU Utilization




                                                          60%

                                                          50%

                                                          40%

                                                          30%

                                                          20%

                                                          10%

                                                           0%
                                                                 1,300 Users   2,200 Users


          Figure 4. Virtual and Physical Host Processor Utilization Comparison

          In the UP configuration at 1,300 users, the average host processor utilization is about 51% for physical and
          about 80% for virtual. In the SMP configuration at 2,200 users, the average host processor utilization is about
          49% for physical and about 80% for virtual. The difference in the host processor utilizations in both cases can
          primarily be attributed to the CPU virtualization overhead.

          Enterprises should keep in mind that the processor virtualization overhead for Exchange is typically not a
          problem in real-world deployments. A survey conducted by VMware and RapidApp of a number of small,
          medium, and large businesses revealed that the weighted average peak CPU utilization for production
          Exchange servers is about 16% (calculated from Figure 10, page 17, of Deploying Microsoft Exchange in VMware
          Infrastructure; see “Resources and References” on page 15 for additional details), leaving substantial headroom
          for virtualization overhead.


     Memory Performance
          With each virtual machine configured with 4GB of RAM, one might expect a four-VM configuration to utilize
          the entire 16GB of physical memory. However Figure 5 shows about 3.5GB of memory remaining in the
          four-VM case. This is due to ESX Server’s transparent page sharing, which shares common memory pages




10                                                                                                             VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



          across virtual machines and can allow over-commitment of virtual machine memory without impact on
          performance. Figure 6 shows this page sharing activity on the ESX Server host across different configurations.


                                  14,000


                                  12,000


                                  10,000
               Free Memory (MB)




                                   8,000


                                   6,000


                                   4,000


                                   2,000


                                      0
                                           One 1-vCPU VM     One 2-vCPU VM    Two 1-vCPU VMs     Two 2-vCPU VMs      Four 1-vCPU VMs


          Figure 5. Host Free Memory

          Figure 6 also plots the ‘swap bytes’ counter from esxtop, which is zero across all configurations. Because there
          is no memory overcommitment, even with four VMs, there is no need for ESX Server swap activity that could
          otherwise have adversely impacted performance.



                                  4,000


                                  3,500

                                                    Page Share Savings (MB)
                                  3,000
                                                    Swap (MB/s)

                                  2,500
                  Megabytes




                                  2,000


                                  1,500


                                  1,000


                                    500


                                      0
                                           One 1-vCPU VM     One 2-vCPU VM    Two 1-vCPU VMs      Two 2-vCPU VMs      Four 1-vCPU VMs



          Figure 6. Page Sharing and Swap Activity for Scale Up and Scale Out

          A single Exchange 2003 instance running in a physical machine is limited to using at most 4GB of physical
          memory. Since Exchange is memory intensive, this can limit the number of Exchange users a 32-bit physical
          system can support. Virtualization’s ability to run multiple instances of these 32-bit operating systems on the
          same physical system effectively removes this 4GB-per-system memory limit, thus increasing the number of
          users a single physical system can support. This constitutes a major advantage of the virtual setup over the
          physical one.



VMware, Inc.                                                                                                                                 11
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



     Networking Performance
          The virtualized configurations generated approximately the same amount of networking traffic as the
          corresponding physical configurations:

          1,300 users on either one physical CPU or one virtual CPU generated:

               Receive traffic: approximately 1 Mbit/s
               Transmit traffic: approximately 2 Mbit/s

          2,200 users on either two physical CPUs or two virtual CPUs generated:

               Receive traffic: approximately 1 Mbit/s
               Transmit traffic: approximately 4 Mbit/s



                             7

                                            Receive
                             6
                                            Transmit


                             5
                Megabits/s




                             4


                             3


                             2


                             1


                             0
                                 One 1-vCPU VM     One 2-vCPU VM    Two 1-vCPU VMs   Two 2-vCPU VMs   Four 1-vCPU VMs


          Figure 7. Network Traffic for Scale Up and Scale Out

          As shown in Figure 7, we observed steady growth in both received and transmitted network traffic with an
          increasing number of virtual CPUs. Even at a high of 4,000 Exchange users (for the two SMP VMs and four UP
          VMs) the cumulative network traffic is less than 10 Mbit/s. This is a small fraction of the throughput the
          underlying gigabit network can handle.


     Storage Performance
          The virtualized configurations generated approximately the same amount of storage I/O as the corresponding
          physical configurations:

          1,300 users on either one physical CPU or one virtual CPU generated:

               Read throughput: approximately 1 MB/s
               Write throughput: approximately 2 MB/s

          2,200 users on either two physical CPUs or two virtual CPUs generated:

               Read throughput: approximately 3 MB/s
               Write throughput: approximately 4 MB/s




12                                                                                                               VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



          As shown in Figure 8, we observed that the storage throughput scaled with the number of virtual CPUs
          (except for an unexplained dip in the read rate for the four 1-vCPU VM case, which we are still investigating).



                             9                                                                                          2,500

                             8              Read (MB/s)

                                            Write (MB/s)
                                                                                                                        2,000
                             7
                                            Commands/s

                             6              Queued Commands




                                                                                                                                Commands/s
               Megabytes/s




                                                                                                                        1,500
                             5

                             4
                                                                                                                        1,000
                             3

                             2
                                                                                                                        500

                             1

                             0                                                                                          0
                                 One 1-vCPU VM   One 2-vCPU VM   Two 1-vCPU VMs   Two 2-vCPU VMs    Four 1-vCPU VMs


          Figure 8. Storage Throughputs for Scale Up and Scale Out

          Even at a high of 4,000 Exchange users (for the two SMP VM and the four UP VM scenarios) the cumulative
          storage I/O is about 14 MB/s (or 112 Mbit/s), which is far from saturating the underlying 2Gb/s SAN Fibre
          Channel link. Absence of ESX Server storage “Queued Commands” reported by esxtop in all of the
          configurations indicates absence of any storage bottlenecks.




VMware, Inc.                                                                                                                                 13
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



Conclusion
          Deploying Microsoft Exchange within virtual machines without knowing what performance to expect can
          make IT departments apprehensive. In this paper we have shown that a single uniprocessor Exchange 2003
          instance running within a virtual machine on ESX Server 3.0.1 can comfortably support up to 1,300 Exchange
          users, and with four uniprocessor instances (or two SMP instances) this can be scaled to 4,000 Exchange users.

          Exchange running on multiple uniprocessor virtual machines or half as many dual processor virtual machines
          shows insignificant performance differences. We thus recommend that the Windows and Exchange licensing
          costs, ease of management, and corporate standards should guide your configuration in this regard.

          This paper also demonstrated the manageable overhead due to virtualization of an Exchange environment. A
          survey of the industry’s physical Exchange deployments in production revealed CPU underutilization, with
          an average peak load of about 16%. In the average and moderately more utilized Exchange deployments
          ample resources for virtualization overhead remain. With VMware virtualization providing features to reduce
          maintenance costs and eliminate downtime, a compelling case for virtualizing Exchange can be made.




14                                                                                                           VMware, Inc.
Microsoft Exchange Server 2003 Performance on VMware ESX Server 3



Resources and References

     VMware
          Deploying Microsoft Exchange in VMware Infrastructure:
          http://www.VMware.com/pdf/exchange_best_practices.pdf

          Performance Tuning Best Practices for ESX Server 3:
          http://www.VMware.com/pdf/vi_performance_tuning.pdf

          Server Configuration Guide:
          http://www.VMware.com/pdf/vi3_server_config.pdf

          SAN Configuration Guide:
          http://www.VMware.com/pdf/vi3_esx_san_cfg.pdf


     Dell
          Dell PowerEdge:
          http://www.dell.com/content/products/features.aspx/pedge_6850

          Dell Resource website on VMware:
          http://www.dell.com/content/topics/global.aspx/alliances/en/vmware?c=us&cs=555&l=en&s=biz

          Dell whitepapers on Microsoft Exchange 2003:
          http://www.dell.com/downloads/global/solutions/Dell_Exchange_SMB.pdf

          http://www.dell.com/downloads/global/power/ps1q05-20040216-Stanford.pdf


     Microsoft
          Microsoft Exchange Server 2003 Deployment Guide:
          http://www.microsoft.com/downloads/details.aspx?familyid=77B6D819-C7B3-42D1-8FBB-FE6339FFA1ED&
          displaylang=en

          Microsoft Exchange Server 2003 Load Simulator (LoadSim) Executable and Documentation:
          http://www.microsoft.com/downloads/details.aspx?FamilyID=92eb2edc-3433-47ca-a5f8-0483c7ddea85&Disp
          layLang=en




VMware, Inc.                                                                                                            15

Más contenido relacionado

La actualidad más candente

HP PolyServe Software for Microsoft SQL Server
HP PolyServe Software for Microsoft SQL ServerHP PolyServe Software for Microsoft SQL Server
HP PolyServe Software for Microsoft SQL Serverwebhostingguy
 
Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...
Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...
Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...ijcsit
 
Vsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscsVsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscsDhymas Mahendra
 
Server consolidation with the Dell PowerEdge M620
Server consolidation with the Dell PowerEdge M620Server consolidation with the Dell PowerEdge M620
Server consolidation with the Dell PowerEdge M620Principled Technologies
 
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Principled Technologies
 
Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...
Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...
Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...Principled Technologies
 
DB2 UDB for z/OS Version 7 - An Overview
DB2 UDB for z/OS Version 7 - An OverviewDB2 UDB for z/OS Version 7 - An Overview
DB2 UDB for z/OS Version 7 - An OverviewCraig Mullins
 
Citrix Xen Desktop Solution White Paper
Citrix Xen Desktop Solution White PaperCitrix Xen Desktop Solution White Paper
Citrix Xen Desktop Solution White PaperReadWriteEnterprise
 
Database performance and memory capacity with the Intel Xeon processor E5-266...
Database performance and memory capacity with the Intel Xeon processor E5-266...Database performance and memory capacity with the Intel Xeon processor E5-266...
Database performance and memory capacity with the Intel Xeon processor E5-266...Principled Technologies
 
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SANVMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SANPrincipled Technologies
 
Database performance with Dell PowerEdge PCIe Express Flash SSDs
Database performance with Dell PowerEdge PCIe Express Flash SSDsDatabase performance with Dell PowerEdge PCIe Express Flash SSDs
Database performance with Dell PowerEdge PCIe Express Flash SSDsPrincipled Technologies
 
MailRetriever for DPM: Quick Start Guide
MailRetriever for DPM: Quick Start GuideMailRetriever for DPM: Quick Start Guide
MailRetriever for DPM: Quick Start GuideMailRetriever
 
Hardware upgrades to improve database, SharePoint, Exchange, and file server ...
Hardware upgrades to improve database, SharePoint, Exchange, and file server ...Hardware upgrades to improve database, SharePoint, Exchange, and file server ...
Hardware upgrades to improve database, SharePoint, Exchange, and file server ...Principled Technologies
 
Virtualized database performance with Dell PowerEdge PCIe Express Flash SSDs
Virtualized database performance with Dell PowerEdge PCIe Express Flash SSDsVirtualized database performance with Dell PowerEdge PCIe Express Flash SSDs
Virtualized database performance with Dell PowerEdge PCIe Express Flash SSDsPrincipled Technologies
 
Vmware documentação tecnica
Vmware documentação tecnicaVmware documentação tecnica
Vmware documentação tecnicaALEXANDRE MEDINA
 
Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...
Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...
Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...Principled Technologies
 
Harness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon Processors
Harness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon ProcessorsHarness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon Processors
Harness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon ProcessorsReadWriteEnterprise
 

La actualidad más candente (20)

Installation Guide
Installation GuideInstallation Guide
Installation Guide
 
HP PolyServe Software for Microsoft SQL Server
HP PolyServe Software for Microsoft SQL ServerHP PolyServe Software for Microsoft SQL Server
HP PolyServe Software for Microsoft SQL Server
 
Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...
Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...
Performance Variations in Profiling Mysql Server on the Xen Platform: Is It X...
 
Vsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscsVsphere esxi-vcenter-server-55-setup-mscs
Vsphere esxi-vcenter-server-55-setup-mscs
 
Server consolidation with the Dell PowerEdge M620
Server consolidation with the Dell PowerEdge M620Server consolidation with the Dell PowerEdge M620
Server consolidation with the Dell PowerEdge M620
 
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...
 
Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...
Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...
Resource balancing comparison: VMware vSphere 6 vs. Red Hat Enterprise Virtua...
 
Oracle 11g R2 Live Part 2
Oracle 11g R2 Live Part 2Oracle 11g R2 Live Part 2
Oracle 11g R2 Live Part 2
 
DB2 UDB for z/OS Version 7 - An Overview
DB2 UDB for z/OS Version 7 - An OverviewDB2 UDB for z/OS Version 7 - An Overview
DB2 UDB for z/OS Version 7 - An Overview
 
Citrix Xen Desktop Solution White Paper
Citrix Xen Desktop Solution White PaperCitrix Xen Desktop Solution White Paper
Citrix Xen Desktop Solution White Paper
 
11g R2 Live Part 1
11g R2 Live Part 111g R2 Live Part 1
11g R2 Live Part 1
 
Database performance and memory capacity with the Intel Xeon processor E5-266...
Database performance and memory capacity with the Intel Xeon processor E5-266...Database performance and memory capacity with the Intel Xeon processor E5-266...
Database performance and memory capacity with the Intel Xeon processor E5-266...
 
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SANVMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
VMmark virtualization performance of Micron Enterprise PCIe SSD-based SAN
 
Database performance with Dell PowerEdge PCIe Express Flash SSDs
Database performance with Dell PowerEdge PCIe Express Flash SSDsDatabase performance with Dell PowerEdge PCIe Express Flash SSDs
Database performance with Dell PowerEdge PCIe Express Flash SSDs
 
MailRetriever for DPM: Quick Start Guide
MailRetriever for DPM: Quick Start GuideMailRetriever for DPM: Quick Start Guide
MailRetriever for DPM: Quick Start Guide
 
Hardware upgrades to improve database, SharePoint, Exchange, and file server ...
Hardware upgrades to improve database, SharePoint, Exchange, and file server ...Hardware upgrades to improve database, SharePoint, Exchange, and file server ...
Hardware upgrades to improve database, SharePoint, Exchange, and file server ...
 
Virtualized database performance with Dell PowerEdge PCIe Express Flash SSDs
Virtualized database performance with Dell PowerEdge PCIe Express Flash SSDsVirtualized database performance with Dell PowerEdge PCIe Express Flash SSDs
Virtualized database performance with Dell PowerEdge PCIe Express Flash SSDs
 
Vmware documentação tecnica
Vmware documentação tecnicaVmware documentação tecnica
Vmware documentação tecnica
 
Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...
Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...
Consolidate SAS 9.4 workloads with Intel Xeon processor E7 v3 and Intel SSD t...
 
Harness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon Processors
Harness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon ProcessorsHarness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon Processors
Harness Enterprise Data with SQL Server 2008 R2 and New Intel Xeon Processors
 

Destacado

3852_wlan_revised
3852_wlan_revised3852_wlan_revised
3852_wlan_revisedBalaji Ravi
 
Presentation1
Presentation1Presentation1
Presentation1148359
 
Quest_Software_Best_Practices_for_Exchange_2007
Quest_Software_Best_Practices_for_Exchange_2007Quest_Software_Best_Practices_for_Exchange_2007
Quest_Software_Best_Practices_for_Exchange_2007Balaji Ravi
 
Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008
Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008
Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008Balaji Ravi
 

Destacado (9)

exch2003
exch2003exch2003
exch2003
 
3852_wlan_revised
3852_wlan_revised3852_wlan_revised
3852_wlan_revised
 
Presentation1
Presentation1Presentation1
Presentation1
 
10307021
1030702110307021
10307021
 
Quest_Software_Best_Practices_for_Exchange_2007
Quest_Software_Best_Practices_for_Exchange_2007Quest_Software_Best_Practices_for_Exchange_2007
Quest_Software_Best_Practices_for_Exchange_2007
 
Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008
Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008
Upgrading_and_Migrating_to_Exchange_Server_2007_and_Windows_2008
 
a
aa
a
 
La te x
La te xLa te x
La te x
 
b
bb
b
 

Similar a Virtualizing_Exchange2003

Esx configuration guide
Esx configuration guideEsx configuration guide
Esx configuration guideNaga Raju N
 
Vsphere 4-partner-training180
Vsphere 4-partner-training180Vsphere 4-partner-training180
Vsphere 4-partner-training180Suresh Kumar
 
VMmark virtualization performance of the Lenovo ThinkServer RD630
VMmark virtualization performance of the Lenovo ThinkServer RD630VMmark virtualization performance of the Lenovo ThinkServer RD630
VMmark virtualization performance of the Lenovo ThinkServer RD630Principled Technologies
 
Virtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL ServerVirtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL Serverwebhostingguy
 
Virtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL ServerVirtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL Serverwebhostingguy
 
Virtualizing SharePoint Components
Virtualizing SharePoint ComponentsVirtualizing SharePoint Components
Virtualizing SharePoint Componentswebhostingguy
 
Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01Clint Edmonson
 
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Principled Technologies
 
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Principled Technologies
 
Component upgrades from Intel and Dell can increase VM density and boost perf...
Component upgrades from Intel and Dell can increase VM density and boost perf...Component upgrades from Intel and Dell can increase VM density and boost perf...
Component upgrades from Intel and Dell can increase VM density and boost perf...Principled Technologies
 
Infraestructure WMWARE
Infraestructure  WMWAREInfraestructure  WMWARE
Infraestructure WMWAREaktivfinger
 
Java Standard Edition 5 Performance
Java Standard Edition 5 PerformanceJava Standard Edition 5 Performance
Java Standard Edition 5 Performancewhite paper
 
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Principled Technologies
 
MVUG- From Zero to Hero on Hyper V R2- Part1
MVUG- From Zero to Hero on Hyper V R2- Part1MVUG- From Zero to Hero on Hyper V R2- Part1
MVUG- From Zero to Hero on Hyper V R2- Part1Lai Yoong Seng
 
Dell Active System 800 converged infrastructure solution: VDI and collaborati...
Dell Active System 800 converged infrastructure solution: VDI and collaborati...Dell Active System 800 converged infrastructure solution: VDI and collaborati...
Dell Active System 800 converged infrastructure solution: VDI and collaborati...Principled Technologies
 
System Center
System CenterSystem Center
System CenterBlauge
 
Ensure greater uptime and boost VMware vSAN cluster performance with the Del...
Ensure greater uptime and boost VMware vSAN cluster  performance with the Del...Ensure greater uptime and boost VMware vSAN cluster  performance with the Del...
Ensure greater uptime and boost VMware vSAN cluster performance with the Del...Principled Technologies
 
Exchange 2010 New England Vmug
Exchange 2010 New England VmugExchange 2010 New England Vmug
Exchange 2010 New England Vmugcsharney
 

Similar a Virtualizing_Exchange2003 (20)

Esx configuration guide
Esx configuration guideEsx configuration guide
Esx configuration guide
 
IBM XIV Gen3 Storage System
IBM XIV Gen3 Storage SystemIBM XIV Gen3 Storage System
IBM XIV Gen3 Storage System
 
Vsphere 4-partner-training180
Vsphere 4-partner-training180Vsphere 4-partner-training180
Vsphere 4-partner-training180
 
VMmark virtualization performance of the Lenovo ThinkServer RD630
VMmark virtualization performance of the Lenovo ThinkServer RD630VMmark virtualization performance of the Lenovo ThinkServer RD630
VMmark virtualization performance of the Lenovo ThinkServer RD630
 
Virtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL ServerVirtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL Server
 
Virtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL ServerVirtual Solution for Microsoft SQL Server
Virtual Solution for Microsoft SQL Server
 
Virtualizing SharePoint Components
Virtualizing SharePoint ComponentsVirtualizing SharePoint Components
Virtualizing SharePoint Components
 
Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01Windows Server 2008 R2 Dev Session 01
Windows Server 2008 R2 Dev Session 01
 
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
 
Power edge mx7000_sds_performance_1018
Power edge mx7000_sds_performance_1018Power edge mx7000_sds_performance_1018
Power edge mx7000_sds_performance_1018
 
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
 
Component upgrades from Intel and Dell can increase VM density and boost perf...
Component upgrades from Intel and Dell can increase VM density and boost perf...Component upgrades from Intel and Dell can increase VM density and boost perf...
Component upgrades from Intel and Dell can increase VM density and boost perf...
 
Infraestructure WMWARE
Infraestructure  WMWAREInfraestructure  WMWARE
Infraestructure WMWARE
 
Java Standard Edition 5 Performance
Java Standard Edition 5 PerformanceJava Standard Edition 5 Performance
Java Standard Edition 5 Performance
 
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
Upgrade to Dell EMC PowerEdge R940 servers with VMware vSphere 7.0 and gain g...
 
MVUG- From Zero to Hero on Hyper V R2- Part1
MVUG- From Zero to Hero on Hyper V R2- Part1MVUG- From Zero to Hero on Hyper V R2- Part1
MVUG- From Zero to Hero on Hyper V R2- Part1
 
Dell Active System 800 converged infrastructure solution: VDI and collaborati...
Dell Active System 800 converged infrastructure solution: VDI and collaborati...Dell Active System 800 converged infrastructure solution: VDI and collaborati...
Dell Active System 800 converged infrastructure solution: VDI and collaborati...
 
System Center
System CenterSystem Center
System Center
 
Ensure greater uptime and boost VMware vSAN cluster performance with the Del...
Ensure greater uptime and boost VMware vSAN cluster  performance with the Del...Ensure greater uptime and boost VMware vSAN cluster  performance with the Del...
Ensure greater uptime and boost VMware vSAN cluster performance with the Del...
 
Exchange 2010 New England Vmug
Exchange 2010 New England VmugExchange 2010 New England Vmug
Exchange 2010 New England Vmug
 

Más de Balaji Ravi

Más de Balaji Ravi (10)

forgot_administrator_password.htm
forgot_administrator_password.htmforgot_administrator_password.htm
forgot_administrator_password.htm
 
lost-xp-password.html
lost-xp-password.htmllost-xp-password.html
lost-xp-password.html
 
1.Routing-eng
1.Routing-eng1.Routing-eng
1.Routing-eng
 
Practical-LDAP-and-Linux
Practical-LDAP-and-LinuxPractical-LDAP-and-Linux
Practical-LDAP-and-Linux
 
bldef_ap.htm
bldef_ap.htmbldef_ap.htm
bldef_ap.htm
 
L03-ajc-C-arrays
L03-ajc-C-arraysL03-ajc-C-arrays
L03-ajc-C-arrays
 
show.php.htm
show.php.htmshow.php.htm
show.php.htm
 
182
182182
182
 
cryptography
cryptographycryptography
cryptography
 
Balaji
BalajiBalaji
Balaji
 

Virtualizing_Exchange2003

  • 1. Microsoft Exchange Server 2003 Performance on VMware® ESX Server 3 VMware ESX Server 3
  • 2. Microsoft Exchange Server 2003 Performance on VMware® ESX Server 3 Microsoft Exchange Server 2003 Performance on VMware® ESX Server 3 Revision: 20070705 Item: ESX-ENG-Q207-388 You can find the most up-to-date technical documentation on our Web site at http://www.vmware.com/support/ The VMware Web site also provides the latest product updates. If you have comments about this documentation, submit your feedback to: docfeedback@vmware.com © 2007 VMware, Inc. All rights reserved. Protected by one or more of U.S. Patent Nos. 6,397,242, 6,496,847, 6,704,925, 6,711,672, 6,725,289, 6,735,601, 6,785,886, 6,789,156, 6,795,966, 6,880,022, 6,944,699, 6,961,806, 6,961,941, 7,069,413, 7,082,598, 7,089,377, 7,111,086, 7,111,145, 7,117,481, 7,149,843 and 7,155,558; patents pending. VMware, the VMware “boxes” logo and design, Virtual SMP and VMotion are registered trademarks or trademarks of VMware, Inc. in the United States and/or other jurisdictions. Dell and PowerEdge are registered trademarks of Dell, Inc. All other marks and names mentioned herein may be trademarks of their respective companies. VMware, Inc. Dell, Inc. 3145 Porter Drive www.dell.com/vmware Palo Alto, CA 94304 www.vmware.com ii VMware, Inc.
  • 3. Contents Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 1 Introduction 1 Experiment Configuration and Methodology 2 Test Bed Configuration 2 Test and Measurement Tools 5 Test Cases and Test Method 5 Experimental Results and Performance Analysis 7 Physical vs. Virtual Performance 7 Scale Up and Scale Out Performance 7 UP or SMP Virtual Machines? 8 Response Time Overheads 8 System Component Performance 10 Processor Performance 10 Memory Performance 10 Networking Performance 12 Storage Performance 12 Conclusion 14 Resources and References 15 VMware 15 Dell 15 Microsoft 15 VMware, Inc. iii
  • 4. Virtualizing Microsoft Exchange 2003 Deployments on VMware ESX Server 3 iv VMware, Inc.
  • 5. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Introduction Virtualization has become a mainstream technology, allowing enterprises to consolidate underutilized servers while helping to increase reliability and fault tolerance and simplify load balancing. As organizations embrace virtualization in the data center, many may consider virtualizing Microsoft Exchange software. This paper suggests how an enterprise-critical messaging application like Microsoft Exchange Server 2003 should be sized and deployed on VMware® ESX Server1 to obtain a satisfactory Quality of Service. Specifically, we examine: The performance implications of running Exchange Server 2003 on a virtual machine versus a physical system. The performance of Exchange Server 2003 in virtual machine configurations when “scaling-up” (adding more processors to a machine) and “scaling-out” (adding more machines). This paper discusses the performance and scalability of Exchange Server 2003 when it is deployed within virtual machines hosted by VMware ESX Server 3.0.1 on a Dell® PowerEdge® 6850 server with a Dell-EMC CX500 FC SAN. The Heavy user profile from Microsoft’s Exchange Server 2003 Load Simulator benchmarking tool was used to simulate the Exchange workload. Results indicated that a uniprocessor virtual machine can support up to 1,300 Heavy users. Our experiments also show that consolidating multiple instances of these uniprocessor Exchange virtual machines on a PowerEdge 6850 can cumulatively support up to 4,000 Heavy users while still providing acceptable performance and scaling. A key observation in the study is that uniprocessor virtual machines are, from a performance perspective, equivalent to half as many multiprocessor (two virtual processors) virtual machines. Hence we recommend that the Windows and Exchange licensing costs, ease of management, and corporate standards guide your configuration in this regard. 1. When considering the deployment of Microsoft Exchange Server 2003 on a virtualization platform such as VMware ESX Server, please refer to http://support.microsoft.com/kb/897615 for information about the Microsoft support policy for their applications running in non-Microsoft virtualization software. VMware, Inc. 1
  • 6. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Experiment Configuration and Methodology The performance and sizing studies were done at the Dell Enterprise Solutions Engineering Labs in collaboration with VMware engineers. The purpose of the tests was to measure, analyze, and understand the performance of Exchange in both the physical and virtual environments. In the following sections the test bed configuration used for the experiments is described in detail, and the test tools are discussed. Finally, we present a description of the experiments. Test Bed Configuration In our test configuration the system under test (SUT) was a Dell PowerEdge 6850 server. The PowerEdge 6850 was configured with four 2.66 GHz dual-core Intel Xeon 7020 processors and 16GB of RAM. In the physical environment, the system was running Microsoft Windows Server 2003 Enterprise Edition (32-bit) and Microsoft Exchange Server 2003 with Service Pack 2 (SP2). In the virtualized environment, the system was running VMware ESX Server 3.0.1, and the virtual machines were configured with the same operating system and application as in the physical system. The client systems were Dell PowerEdge 1855 blade servers, each with two 3.6 GHz Intel Xeon processors and 2GB of RAM. One client system was deployed for each Exchange mailbox store. The client systems were running Microsoft Windows Server 2003 Enterprise Edition (32-bit) and Microsoft Exchange Server 2003 Load Simulator (LoadSim). In addition to its function as a mail client, the first of these client systems also served as the active directory server, the domain controller, and the DNS server. Table 1 through Table 3 summarize the test setup in more detail. Table 1. System Under Test (SUT) Configuration Server Dell PowerEdge 6850 Processors Four 2.66 GHz dual-core Intel Xeon 7020 (eight total cores) Memory 16GB DDR-2 400 ECC SDRAM Hard drives (for operating system) Two 146GB 10,000 RPM drives in a RAID 1 array HBA Two Emulex Fibre Channel HBAs Virtualization software VMware ESX Server 3.0.1 Operating system (physical and virtual) Microsoft Windows Server 2003 Enterprise Edition (32-bit) Application (physical and virtual) Microsoft Exchange Server 2003 with Service Pack 2 Table 2. Storage Configuration Storage enclosure Dell EMC CX500 with one disk processor enclosure (DPE) and three disk array enclosures (DAE) Hard drives Fifty-four 73GB 15,000 RPM Drives RAID configuration Eight 6-drive RAID 1+0 volumes for Microsoft Exchange Information Store (IS) One 6-drive RAID 1+0 volume partitioned into eight LUNs for the transaction log 2 VMware, Inc.
  • 7. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Table 3. Client Configuration Client Dell PowerEdge 1855 with four blade servers Processors Two 3.6 GHz Intel Xeon per blade Memory 2GB per blade Operating system Microsoft Windows Server 2003 Enterprise Edition (32-bit) Benchmark application Microsoft Exchange Server 2003 Load Simulator (LoadSim) VMware, Inc. 3
  • 8. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Client (Dell® PowerEdge® 1855 blade servers) Gigabit Ethernet switch Server (Dell® PowerEdge® 6850) Fibre Channel switch (Brocade Silkworm 3900) Storage (Dell/EMC CX500) Figure 1. Test Configuration Diagram 4 VMware, Inc.
  • 9. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Test and Measurement Tools Microsoft LoadSim 2003, running on the client systems, was used to simulate messaging load for the physical and virtual Exchange deployments. LoadSim provided the tools to measure how a system would respond when supporting a given number of Exchange users, in this case Messaging Application Programming Interface (MAPI) clients simulated using the LoadSim Heavy user profile. For a LoadSim run to be valid, 95% of the response times experienced by these users must be less than 500ms. This ensures that the end users get reasonable interactive performance. The benchmark tries to closely model the normal daily email usage of real users in order to provide an estimate of the number of users a system can support. While this benchmark is widely used to measure the performance of Exchange platforms, as with all benchmarks the results may not match the specifics of your environment. We used Windows Perfmon, a performance monitoring tool for Microsoft Windows operating systems, to monitor performance on the physical Exchange setups. We configured Perfmon to log relevant CPU, memory, disk, network, and system counters as well as Exchange-specific counters. These counters were logged to a comma-separated value (CSV) performance log which was later imported to a spreadsheet for analysis. We used esxtop, a performance monitoring tool for ESX Server, to record both ESX Server and virtual machine related performance counters for the virtualized test cases. We configured esxtop to log processor, memory, disk, network, and system counters during the LoadSim runs. The data from Perfmon and esxtop are the primary focus of our analysis. Test Cases and Test Method We had two primary objectives in performing these experiments: Comparing the performance of physical Exchange installations with their virtual counterparts. Understanding the scalability of the virtual machines (VMs) with respect to the number of VMs and the Virtual CPU count. We therefore created the following test cases: Physical uniprocessor (UP physical) Physical symmetric multiprocessor with two processors (SMP physical) One VM with one virtual CPU (UP VM) Two UP VMs Four UP VMs One VM with two virtual CPUs (SMP VM) Two SMP VMs As shown in Figure 4, these configurations can be compared in several ways to understand the virtualization overhead, “scale up” potential, and “scale out” potential. Table 4. Test Cases Comparison Measurement UP physical vs. UP VM Virtualization overhead SMP physical vs. SMP VM Virtualization overhead UP VM vs. SMP VM Scale up potential One UP VM vs. Two UP VMs Scale out potential One UP VM vs. Four UP VMs Scale out potential One SMP VM vs. Two SMP VMs Scale out potential VMware, Inc. 5
  • 10. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Our goal for each configuration was to find the number of LoadSim Heavy users that the configuration could support while continuing to meet the following Quality of Service (QoS) metrics: The host system CPU utilization must average 80%. This matches the typical maximum load in production environments and enables benchmark score comparison across configurations. 95% of the response times must be below 500ms (as required for a LoadSim run to be valid). Each test was performed twice in order to verify the consistency and repeatability of the results. To benchmark the UP and SMP physical configurations we booted the 8-way physical box with one and two processors, respectively, and with 4GB of memory. This was achieved by appropriately setting the parameters /NUMPROC and /MAXMEM in the Windows C:boot.ini system file. Being a 32-bit application that doesn't support Physical Address Extension (PAE), a single instance of Exchange Server 2003 is limited to using at most 4GB of physical memory. ESX Server was configured to use the entire 16GB of memory available on the system, but each VM running on ESX Server was configured to use 4GB to match the physical configuration. In both the physical and virtual environments the /PAE and /3GB options in the C:boot.ini system file were used, the later option allowing Exchange to address up to 3GB of memory. 6 VMware, Inc.
  • 11. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Experimental Results and Performance Analysis In this section we discuss the results obtained from the experiments described above. First we present the measurements for comparable physical and virtual systems, then we discuss the scale up and scale out performance as seen in virtual environments. Physical vs. Virtual Performance In order to compare the performance of physical and virtual environments, we conducted UP and SMP LoadSim experiments, first on the physical system, then in the ESX Server 3.0.1 virtual environment. The software stacks, Windows Server 2003 running Exchange Server 2003, were configured identically on the physical and virtual systems. Doing so enabled an “apples to apples” comparison between the physical and virtual setups. The results of these experiments are shown in Figure 2. 3,500 Virtual 3,000 Physical LoadSim Heavy Users 2,500 2,000 1,500 1,000 500 0 One Processor Two Processors Figure 2. UP and SMP Physical and Virtual Throughput Comparisons In the UP case we recorded scores of 1,300 and 1,700 users for the virtual and physical setups, respectively, which is a virtual to physical ratio of about 0.76. In the SMP case the virtual and physical environments managed to support 2,200 and 3,100 users, respectively, which is a virtual to physical ratio of about 0.71. Scale Up and Scale Out Performance In addition to comparing the virtual to physical performance, we also wanted to study the scale out performance (by increasing the number of VMs) and the scale up performance (by increasing the number of virtual CPUs configured for each VM). We conducted experiments using the following configurations: Scale out (increasing the number of VMs): One, two, and four uniprocessor virtual machines. One and two dual-processor virtual machines. Scale up (increasing the number of processors per VM): Uniprocessor and dual-processor virtual machines. VMware, Inc. 7
  • 12. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 The results of these scale up and scale out experiments are shown in Figure 3 and detailed in Table 5. 4,500 4,000 3,500 Number of Users 3,000 2,500 2,000 1,500 1,000 500 0 One 1-vCPU VM One 2-vCPU VM Two 1-vCPU VMs Two 2-vCPU VMs Four 1-vCPU VMs Figure 3. Virtual Scale Out and Scale Up Performance Table 5. Scale Out and Scale Up Results Experiment First Data Point Second Data Point Scaling Uniprocessor scale out One UP VM (1,300 users) Two UP VMs (2,400 users) 1.85 Uniprocessor scale out One UP VM (1,300 users) Four UP VMs (4,000 users) 3.08 Dual processor scale out One SMP VM (2,200 users) Two SMP VMs (4,000 users) 1.82 Processor scale up One UP VM (1,300 users) One SMP VM (2,200 users) 1.69 UP or SMP Virtual Machines? A fixed number of virtual CPUs (vCPUs) can be deployed in a number of ways. For instance, two VMs – each with one vCPU – might be compared to a single VM with two vCPUs. Each scenario has the same number of vCPUs, but may experience different performance depending on application characteristics. Two VMs configured with one vCPU each generated a score of 2,400 LoadSim Heavy users, while one VM with two vCPUs scored 2,200 LoadSim Heavy users. This might suggest that configuring multiple UP VMs improves performance compared to fewer SMP VMs. However the scenario with four uniprocessor VMs and the scenario with two dual-processor VMs each generated a score of 4,000 LoadSim Heavy users. So both the UP and SMP configurations performed approximately equally in our experiments. It should be noted that considerations such as licensing costs for the different configurations may result in different price-performance recommendations. In addition to the software licensing costs, the ease of management (with fewer VMs) or corporate standards (standardizing on a UP or SMP image) might dictate the choice of UP VMs or half as many SMP VMs. Response Time Overheads The CPU, memory, networking, and storage virtualization overheads induce higher response times for the Exchange benchmark running within the virtual machines compared to the physical setup. A fair comparison 8 VMware, Inc.
  • 13. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 using LoadSim with the same number of Exchange users both physically and within the VMs results in the 95th percentile response times shown in Table 6. Table 6. Response Time Overheads Virtual Physical One processor (1,300 users) 365ms 292ms Two processors (2,200 users) 440ms 260ms The data shows an increase in the response time for the virtual setup over the physical one. But it bears repeating that even with an increased latency of around 100-200ms, the virtual environments still met the 500ms QoS requirement. VMware, Inc. 9
  • 14. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 System Component Performance The previous section discussed the benchmark throughput across the various virtual and physical configurations of interest. This section will analyze the impact of those throughputs on the system components including the processors, memory, network, and storage. Processor Performance As previously mentioned, all of the physical and virtual configurations were pushed to 80% average host processor utilization and the LoadSim throughputs at that CPU limit were measured. The difference in the virtual and physical scores can be attributed to the processor virtualization overheads combined with memory, storage, and networking overheads. To get a clearer understanding of just the processor virtualization overhead we re-ran the physical UP and physical SMP experiments with the same number of users as the maximum we had achieved with the UP VM and SMP VM virtual environments (1,300 and 2,200 users, respectively). The host processor utilizations for both the virtual and physical configurations are shown in Figure 4. 100% Physical 90% Virtual 80% 70% CPU Utilization 60% 50% 40% 30% 20% 10% 0% 1,300 Users 2,200 Users Figure 4. Virtual and Physical Host Processor Utilization Comparison In the UP configuration at 1,300 users, the average host processor utilization is about 51% for physical and about 80% for virtual. In the SMP configuration at 2,200 users, the average host processor utilization is about 49% for physical and about 80% for virtual. The difference in the host processor utilizations in both cases can primarily be attributed to the CPU virtualization overhead. Enterprises should keep in mind that the processor virtualization overhead for Exchange is typically not a problem in real-world deployments. A survey conducted by VMware and RapidApp of a number of small, medium, and large businesses revealed that the weighted average peak CPU utilization for production Exchange servers is about 16% (calculated from Figure 10, page 17, of Deploying Microsoft Exchange in VMware Infrastructure; see “Resources and References” on page 15 for additional details), leaving substantial headroom for virtualization overhead. Memory Performance With each virtual machine configured with 4GB of RAM, one might expect a four-VM configuration to utilize the entire 16GB of physical memory. However Figure 5 shows about 3.5GB of memory remaining in the four-VM case. This is due to ESX Server’s transparent page sharing, which shares common memory pages 10 VMware, Inc.
  • 15. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 across virtual machines and can allow over-commitment of virtual machine memory without impact on performance. Figure 6 shows this page sharing activity on the ESX Server host across different configurations. 14,000 12,000 10,000 Free Memory (MB) 8,000 6,000 4,000 2,000 0 One 1-vCPU VM One 2-vCPU VM Two 1-vCPU VMs Two 2-vCPU VMs Four 1-vCPU VMs Figure 5. Host Free Memory Figure 6 also plots the ‘swap bytes’ counter from esxtop, which is zero across all configurations. Because there is no memory overcommitment, even with four VMs, there is no need for ESX Server swap activity that could otherwise have adversely impacted performance. 4,000 3,500 Page Share Savings (MB) 3,000 Swap (MB/s) 2,500 Megabytes 2,000 1,500 1,000 500 0 One 1-vCPU VM One 2-vCPU VM Two 1-vCPU VMs Two 2-vCPU VMs Four 1-vCPU VMs Figure 6. Page Sharing and Swap Activity for Scale Up and Scale Out A single Exchange 2003 instance running in a physical machine is limited to using at most 4GB of physical memory. Since Exchange is memory intensive, this can limit the number of Exchange users a 32-bit physical system can support. Virtualization’s ability to run multiple instances of these 32-bit operating systems on the same physical system effectively removes this 4GB-per-system memory limit, thus increasing the number of users a single physical system can support. This constitutes a major advantage of the virtual setup over the physical one. VMware, Inc. 11
  • 16. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Networking Performance The virtualized configurations generated approximately the same amount of networking traffic as the corresponding physical configurations: 1,300 users on either one physical CPU or one virtual CPU generated: Receive traffic: approximately 1 Mbit/s Transmit traffic: approximately 2 Mbit/s 2,200 users on either two physical CPUs or two virtual CPUs generated: Receive traffic: approximately 1 Mbit/s Transmit traffic: approximately 4 Mbit/s 7 Receive 6 Transmit 5 Megabits/s 4 3 2 1 0 One 1-vCPU VM One 2-vCPU VM Two 1-vCPU VMs Two 2-vCPU VMs Four 1-vCPU VMs Figure 7. Network Traffic for Scale Up and Scale Out As shown in Figure 7, we observed steady growth in both received and transmitted network traffic with an increasing number of virtual CPUs. Even at a high of 4,000 Exchange users (for the two SMP VMs and four UP VMs) the cumulative network traffic is less than 10 Mbit/s. This is a small fraction of the throughput the underlying gigabit network can handle. Storage Performance The virtualized configurations generated approximately the same amount of storage I/O as the corresponding physical configurations: 1,300 users on either one physical CPU or one virtual CPU generated: Read throughput: approximately 1 MB/s Write throughput: approximately 2 MB/s 2,200 users on either two physical CPUs or two virtual CPUs generated: Read throughput: approximately 3 MB/s Write throughput: approximately 4 MB/s 12 VMware, Inc.
  • 17. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 As shown in Figure 8, we observed that the storage throughput scaled with the number of virtual CPUs (except for an unexplained dip in the read rate for the four 1-vCPU VM case, which we are still investigating). 9 2,500 8 Read (MB/s) Write (MB/s) 2,000 7 Commands/s 6 Queued Commands Commands/s Megabytes/s 1,500 5 4 1,000 3 2 500 1 0 0 One 1-vCPU VM One 2-vCPU VM Two 1-vCPU VMs Two 2-vCPU VMs Four 1-vCPU VMs Figure 8. Storage Throughputs for Scale Up and Scale Out Even at a high of 4,000 Exchange users (for the two SMP VM and the four UP VM scenarios) the cumulative storage I/O is about 14 MB/s (or 112 Mbit/s), which is far from saturating the underlying 2Gb/s SAN Fibre Channel link. Absence of ESX Server storage “Queued Commands” reported by esxtop in all of the configurations indicates absence of any storage bottlenecks. VMware, Inc. 13
  • 18. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Conclusion Deploying Microsoft Exchange within virtual machines without knowing what performance to expect can make IT departments apprehensive. In this paper we have shown that a single uniprocessor Exchange 2003 instance running within a virtual machine on ESX Server 3.0.1 can comfortably support up to 1,300 Exchange users, and with four uniprocessor instances (or two SMP instances) this can be scaled to 4,000 Exchange users. Exchange running on multiple uniprocessor virtual machines or half as many dual processor virtual machines shows insignificant performance differences. We thus recommend that the Windows and Exchange licensing costs, ease of management, and corporate standards should guide your configuration in this regard. This paper also demonstrated the manageable overhead due to virtualization of an Exchange environment. A survey of the industry’s physical Exchange deployments in production revealed CPU underutilization, with an average peak load of about 16%. In the average and moderately more utilized Exchange deployments ample resources for virtualization overhead remain. With VMware virtualization providing features to reduce maintenance costs and eliminate downtime, a compelling case for virtualizing Exchange can be made. 14 VMware, Inc.
  • 19. Microsoft Exchange Server 2003 Performance on VMware ESX Server 3 Resources and References VMware Deploying Microsoft Exchange in VMware Infrastructure: http://www.VMware.com/pdf/exchange_best_practices.pdf Performance Tuning Best Practices for ESX Server 3: http://www.VMware.com/pdf/vi_performance_tuning.pdf Server Configuration Guide: http://www.VMware.com/pdf/vi3_server_config.pdf SAN Configuration Guide: http://www.VMware.com/pdf/vi3_esx_san_cfg.pdf Dell Dell PowerEdge: http://www.dell.com/content/products/features.aspx/pedge_6850 Dell Resource website on VMware: http://www.dell.com/content/topics/global.aspx/alliances/en/vmware?c=us&cs=555&l=en&s=biz Dell whitepapers on Microsoft Exchange 2003: http://www.dell.com/downloads/global/solutions/Dell_Exchange_SMB.pdf http://www.dell.com/downloads/global/power/ps1q05-20040216-Stanford.pdf Microsoft Microsoft Exchange Server 2003 Deployment Guide: http://www.microsoft.com/downloads/details.aspx?familyid=77B6D819-C7B3-42D1-8FBB-FE6339FFA1ED& displaylang=en Microsoft Exchange Server 2003 Load Simulator (LoadSim) Executable and Documentation: http://www.microsoft.com/downloads/details.aspx?FamilyID=92eb2edc-3433-47ca-a5f8-0483c7ddea85&Disp layLang=en VMware, Inc. 15