SlideShare una empresa de Scribd logo
1 de 73
vStorage: Troubleshooting Performance
 Nathan Small, Staff Engineer, Global Support Services




Confidential
                                                         © 2009 VMware Inc. All rights reserved
Agenda
       vCenter Performance Charts & ESXTop
       Disk Alignment
       Hard Disks/Storage Array
       SCSI Reservations
       APD (All Paths Down) Issue
       Multipathing Considerations
       Communication




2                                  Confidential
vCenter Performance Charts & ESXTop




3                   Confidential
Performance Charts

     Through the Overview section of the Performance tab of an ESX host,
     we can observe if a host is seeing recent disk latency issues.




     Unfortunately, this chart does not pinpoint which device or path is
     seeing this latency. For a more granularity, we will use “Advanced”

4                                   Confidential
Performance Charts

     If we switch to the “Storage path” option, we can see latency is
     coming from vmhba0:C0:T0:L0 for both read and write latency:




5                                Confidential
Performance Charts

     In order to see the latency statistics for a given LUN, we will need
     to go to the “Disk” view and then select “Chart Options”:




    Select the LUNs you wish to
    see the latency statistics for
    under “Objects”, then select
    the Physical device read and
    write latency “Counters”




6                                    Confidential
Performance Charts

     With this granular view, we can see that the LUN with ID
     naa.60060480000190300919533030303134 had a latency spike up
     to 48ms at ~9:50am




7                             Confidential
ESXTop

     ESXTop is a great tool to use if you need performance captured
     during a specific timeframe (scripted batch mode) or if the ESX
     host is unavailable in vCenter/VI Client.
     To view the equivalent of the “Storage path” view, hit „d‟:




8                                   Confidential
ESXTop

     You can toggle the display fields by hitting „f‟. By taking out
     unneeded fields (I/O stats), you can make the display screen far
     easier to read and work with:




9                                Confidential
ESXTop

      To view individual LUN statistics (with unique ID), hit „u‟:




      As we can see, the default field size for “DEVICE” is not long
         enough to show the entire unique identifier. To change this field
         size, hit „L‟:
     




10                                   Confidential
ESXTop

      Expanding this field now shows the entire unique ID for the LUN
      and NFS datastores:




      Just as before, we should display only the fields we need to be
      concerned with by hitting „f‟:




11                                     Confidential
ESXTop

      This view now shows us similar output to the „d‟ view except we
      see the statistics for a LUN (all paths).




     *Note: The reason there are no statistics for the NFS mounts is due
      to the fact that these are ISO and template repositories that are not
      in use currently by this host.




12                                   Confidential
ESXTop

      To capture esxtop data, use „vm-support -S‟. Here is an example
      for collecting these statistics at 2 second intervals over 20
      seconds:
     # vm-support –S –i 2 –d 20




13                                Confidential
ESXTop

      To replay that capture data, extract the compressed file, change to
      the „snaphots‟ directory, run the uncompress script, then run
      „esxtop‟ in replay mode „-R‟:




14                                  Confidential
DAVG/KAVG/GAVG/QAVG

      What is the DAVG?
      • The DAVG field is the Device Average, which is the amount of time it takes
        for a SCSI command to leave the HBA, hit the array, and return completed.
        This is the most effective method of determining if the performance issue
        being experienced is at the physical layer (switch/array).
      What are acceptable DAVG values?
      • Optimal, sustained DAVG latency values would be between 0-5ms for 4/8GB
        FC and 10GB iSCSI/NFS. Seeing 5-10ms latency for 1/2GB FC or 1GB
        iSCSI/NFS is also acceptable.
      • Seeing latency values of 10-20ms could start to show some minor
        performance issues inside the VM. while values of 20-50ms would show
        noticeable to significant performance issues. Latency values of 50ms or
        higher would make the VMs almost unusable.
      How should I proceed when my DAVG is sustained at a high level?
      • Engage the storage team immediately. They may already be aware of the
        latency due to increased load on the storage array (LUN replication, backups)

15                                      Confidential
DAVG/KAVG/GAVG/QAVG

      What is the KAVG?
      • The KAVG field is the Kernel Average, which is the amount of time a SCSI
        command spends in the vmkernel.
      What are acceptable KAVG values?
      • The KAVG should always be less than 1ms.
      How should I proceed when my KAVG is sustained at a high level?
      • While rare to see the KAVG value high at all, it has been observed when the
        queue has been throttled back on the ESX host due to a TASK_SET_FULL
        condition on the array (KB 1008113), or the queue depth has been reduced
        on the HBAs due to configuration restrictions (KB 1006001). It has also been
        observed temporarily on path failover. This is due to commands queuing
        during the path activation process.




16                                      Confidential
DAVG/KAVG/GAVG/QAVG

      What is the GAVG?
      • The GAVG field is the Device Average + Kernel Average + Queue Average,
        which is the latency perceived by the Guest OS or VM.
      What are acceptable GAVG values?
      • As the KAVG and QAVG should always be less than 1ms, acceptable GAVG
        latency values would be the same as acceptable DAVG values.
      How should I proceed when my GAVG is sustained at a high
      level?
       • The same steps should be followed for high DAVG values.




17                                     Confidential
DAVG/KAVG/GAVG/QAVG

      What is the QAVG?
      • The QAVG field is the Queue Average, which is the amount of time the SCSI
        command spends in the HBA driver.
      What are acceptable QAVG values?
      • The QAVG should always be less than 1ms.
      How should I proceed when my QAVG is sustained at a high
      level?
       • Just like the KAVG, the QAVG should not have a high latency value, unless
        there is a legitimate reason for commands to remained queued on the host
        side.




18                                     Confidential
Queue Depth and ESXTop

      Many people believe that increasing the queue depth will solve a
      performance issue or improve performance overall, however this
      can have the opposite effect (KB 1006001 & 1008113).
      The queue depth should only be increased if the array vendor
      recommends to do so or the queue depth is getting exhausted on
      the host side.
      As seen in a previous slide, we can observe the queue depth
      usage and activity:




19                                Confidential
Agenda
        vCenter Performance Charts & ESXTop
        Disk Alignment
        Hard Disks/Storage Array
        SCSI Reservations
        APD (All Paths Down) Issue
        Multipathing Considerations
        Communication




20                                  Confidential
Disk Alignment




21        Confidential
Disk Alignment

      The concept of Disk Alignment is simple, yet the performance hit
      for misaligned I/O can be troublesome for a single, high disk I/O
      VM (SQL, Exchange, etc) or for the entire virtualized infrastructure
      on a particular storage array.
      It has been a best practice recommendation from high I/O
      application vendors to ensure that the data drive for the
      application is aligned to ensure the best I/O possible, as this will
      avoid any partial read or write operation. In many cases, this data
      drive would be a separate LUN either on local storage or on a
      share storage array.




22                                 Confidential
Disk Alignment

      In a SAN environment, the smallest hardware unit used by a SAN
      storage array to build a LUN out of multiple physical disks is a
      called a chunk or a stripe. To optimize I/O, chunks are usually
      much larger than sectors. Thus a SCSI I/O request that intends to
      read a sector in reality reads one chunk.
      On top of this, in a Windows environment NTFS is formatted in
      blocks ranging from 1MB to 8MB. The file system used by the
      guest operating system optimizes I/O by grouping sectors into so-
      called clusters (allocation units).
      Older operating systems on x86 architectures create partitions
      with a master boot record (MBR) that consumes 63 sectors. This is
      due to legacy BIOS code from the PC that used cylinder, head, and
      sector addressing instead of logical block addressing (LBA).
      Without LBA, the first track is reserved for the boot code, and the
      first partition starts at cylinder 0, head 1, and sector 1. This is LBA
      63 and is therefore unaligned.

23                                   Confidential
Disk Alignment

      An unaligned partition results in a track crossing and an
      additional I/O, incurring a penalty on latency and throughput. The
      additional I/O (especially if small) can impact system resources
      significantly on some host types. An aligned partition ensures that
      the single I/O is serviced by a single device, eliminating the
      additional I/O and resulting in overall performance improvement.
      Unaligned structure may cause many additional I/O operations
      when only one cluster is ready by the guest operating system.




24                                  Confidential
Disk Alignment

      Beginning with ESX 3.0.x, or rather VMFS3, any newly created
      VMFS LUN will be aligned to a starting sector of 128, which is 64k
      (128 * 512 = 65536). As most arrays perform reads/writes in 4k, or a
      value divisible by 4k, this means that the VMFS partition is aligned
      with the underlying storage. When the partition for the VM is
      aligned in the same manner, I/O is aligned across the board.
      I/O improvements on a properly aligned Windows NTFS volume in
      a VMDK on a SAN LUN:




25                                 Confidential
Disk Alignment

      The result of aligning I/O can have significant, measurable
      performance improvements for high disk I/O VMs:


      VMW

      SQL




26                                  Confidential
Disk Alignment and NFS

      Does Disk Alignment only apply to VMFS and not NFS?
      While the NFS volume does not have an additional VMFS layer to
      align, the NFS datastore is aligned to a 4k divisible offset when
      created and the array controllers still operate in 4k read/write
      operations.
      As such, there is still a need to align the OS and data drives of
      VMs residing on an NFS volume.




27                                   Confidential
Verifying Disk Alignment

      How can I tell if the disks inside my VM‟s are aligned?




      This is the default partition offset of 63: 63 * 512 = 32256
28                                    Confidential
Aligning Disks

      To align newly created disk (data disks), simply use diskpart (Windows) or
      fdisk (Linux). For more information, see:
       http://www.vmware.com/pdf/esx3_partition_align.pdf
       http://media.netapp.com/documents/tr-3747.pdf
      Aligning existing disks gets tricky since the file system and data has already
      been written. To simply change the offset of these existing partitions with
      diskpart or fdisk will render them useless.

      vSphere/ESX currently has no way of aligning disks however 3rd party
      vendors have created utilities to do this:
       • Netapp provides a utility called ‘mbrscan’, which will scan your VMs’ disks to see if they are
         aligned, and ‘mbralign’, which will align the disks of a VM (powered off VMs only).

       • Vizioncore makes a product called vOptimizer that will align VM’s disks to 64k offset.
         Alternatively, their vConverter 5.0 product will convert physical or virtual machines and in the
         process align the disk offset to 64k
       • Paragon also makes a utility called Paragon Alignment Tool (PAT). This also comes bundled
         with their flagship product.

29                                                Confidential
Planning Ahead

      The best way to plan ahead with Disk Alignment is to create
      templates for every OS you wish to deploy and align those before
      installing the OS. This way you can deploy VM‟s that will already
      be aligned.




30                                 Confidential
Agenda
        vCenter Performance Charts & ESXTop
        Disk Alignment
        Hard Disks/Storage Array
        SCSI Reservations
        APD (All Paths Down) Issue
        Multipathing Considerations
        Communication




31                                  Confidential
Hard Disks/Storage Array




32             Confidential
SCSI vs. SATA

      The argument of which hard drive technology to use is a well
      known discussion however the answer on which drives to use
      ultimately depends on the application or need.
      SCSI drives are more expensive than SATA drives however this is
      due to the design and performance achieved with this technology.
      SCSI uses a tagged command queuing implementation which
      allows many commands to be outstanding. This design provides
      significant performance gain for the drives/controllers to be able to
      order these commands in most optimal execution manner
      possible.
      SCSI drives also use a processor for executing commands and
      handling the interface while a separate processor handles the
      head positioning through servos.
      SCSI can push a maximum of 640Meg/second


33                                  Confidential
SCSI vs. SATA

      In the early years, SATA used Task Command Queuing (TCQ).
      This queuing technology was designed to help bridge the gap
      between SCSI and PATA/SATA drives however there was too
      much overhead and TCQ simply wasn‟t efficient enough. TCQ is
      sometimes referred to as interrupt queuing.
      SATA 2.0 introduced Native Command Queuing (NCQ), which is
      very similar to the outstanding request queuing that SCSI uses.
      This technology drastically reduces the amount of interrupts that
      TCQ uses due to the integrated first party DMA engine as well as
      the intelligent ordering of commands and interrupting when most
      it is most optimal to do so.
      Due to NCQ and the throughput increases with the SATA 2.0
      standard, SATA can perform well but not to the same level as
      SCSI.
      SATA 3GB/s can push a maximum of 300MB/s

34                                 Confidential
When To Use SCSI

      SCSI drives are by far the most expensive choice, however you
      get what you pay for: Performance.
      There are no limitations for SCSI drives usage however budgets
      are rarely unlimited so SCSI drives are usually purchased for
      production servers/applications (SQL, MS Exchange, etc) or tier
      1/2 storage.
      Because of the throughput capabilities of SCSI, these drives will
      handle I/O bursts far better than SATA, which would rely heavily
      on read/write cache to rough the storm.
      SCSI drives are recommended for use with VMware View
      environments for linked clone and fully deployed desktops. More
      on this later.




35                                  Confidential
When To Use SATA

      While SATA drives are far cheaper than SCSI drives and usually
      higher in disk capacity, they simply do not perform at the same
      level as SCSI drives.
      Installing more write cache in the storage array can assist SATA
      drives through heavy I/O periods however that would make write
      cache the last line of defence to an I/O storm or increasing I/O load
      over time. It may only be a matter of time before write cache fills
      up and then begins the battle of destaging.
      If SATA disks must be used, a configuration of RAID 10 for those
      disks is strongly recommended. This will not guarantee availability
      in the heaviest of I/O storms, but it is the most optimal RAID
      configuration while maintaining redundancy.




36                                  Confidential
Agenda
        vCenter Performance Charts & ESXTop
        Disk Alignment
        Hard Disks/Storage Array
        SCSI Reservations
        APD (All Paths Down) Issue
        Multipathing Considerations
        Communication




37                                  Confidential
SCSI Reservations




38          Confidential
What are SCSI Reservations?

      SCSI Reservations refer to the act of an initiator to using
      SCSI-2 commands 0x16 (RESERVE) and 0x17 (RELEASE) to
      lock a LUN for a specific operation.
      While a reservation is placed on a LUN, if another initiator
      attempts to perform any command on the reserved LUN
      other than an INQUIRY, REQUEST SENSE, or a
      PREVENT/ALLOW MEDIUM REMOVAL command the
      command shall be rejected with RESERVATION CONFLICT
      device status.
      A reservation may only be released by the initiator that
      placed the reservation. A release command sent by another
      initiator will be ignored.
      Any initiator can clear the reservation by issuing a “BUS
      DEVICE RESET” command or a hard “RESET”.



39                                Confidential
SCSI Reservations

      Since VMFS-3 is basically a clustered file system, allowing
      simultaneous access from multiple ESX servers, we need a
      way of preserving the integrity of the file system when more
      than one host is updating it.
      VMFS-3 implements a locking protocol to prevent VMware
      cluster-aware applications from powering on (or otherwise
      sharing) the contents of a given virtual disk on more than
      one host at any given time.
      Part of this locking protocol is based on the notion of on-
      disk locks that protect the metadata on the volume.
      An ESX/ESXi host interested in locked access to metadata
      must atomically check the owner and lock-state fields in the
      lock, and if the lock is free, acquire it by writing out its own
      ID and lock state into the owner and lock-state fields.



40                                Confidential
SCSI Reservations

      The lock manager currently uses SCSI reservations to
      check the lock information, establish itself as the owner by
      writing to the relevant lock fields on disk if the lock is free,
      and then releases the SCSI reservation.
      ESX uses SCSI-2 non-persistent reservations which implies
      that only a single host can reserve the LUN in question at
      any one time. A reboot of that host will clear the reservation
      on the LUN (non-persistent).
      ESX 4.0 introduces limited support for SCSI-3 persistent
      reservations to work with Microsoft Windows 2008
      Clustering, however the LSISAS controller must be used for
      the VMs. Only Windows 2k8 is supported with this
      controller.
      VMFS lock operations are still implemented using SCSI-2
      non-persistent reservations, even in ESX 4.x.


41                                Confidential
A typical SCSI Reservation Conflict message
     vmkernel: 0:12:44:32.598 cpu4:1046)SCSI: vm 1046: 5509: Sync CR
      at 64
     vmkernel: 0:12:44:33.520 cpu4:1046)SCSI: vm 1046: 5509: Sync CR
      at 48
     vmkernel: 0:12:44:34.512 cpu4:1046)SCSI: vm 1046: 5509: Sync CR
      at 32
     vmkernel: 0:12:44:35.482 cpu4:1046)SCSI: vm 1046: 5509: Sync CR
      at 16
     vmkernel: 0:12:44:36.490 cpu2:1046)SCSI: vm 1046: 5509: Sync CR
      at 0
     vmkernel: 0:12:44:36.490 cpu2:1046)WARNING: SCSI: 5519: Failing
      I/O due to too many reservation conflicts
     vmkernel: 0:12:44:36.490 cpu2:1046)WARNING: SCSI: 5615: status
      SCSI reservation conflict, rstatus 0xc0de01 for vmhba2:0:3.
      residual R 919, CR 0, ER 3
     vmkernel: 0:12:44:36.490 cpu2:1046)FSS: 343: Failed with status
      0xbad0022 for f530 28 2 45378334 5f8825b2 17007a7d 1d624ca4 4
      1 0 0 0 0 0

 "CR" stands for conflict retry. In this case an I/O is being
 retried due to reservation conflicts. The number at the end of
 the log statement is the number of retries left. In this case, all
 retries were exhausted so the I/O failed.
42                               Confidential
What causes a SCSI Reservation?

      Administrative operations, such as creating or deleting a virtual disk,
      extending a VMFS volume, or creating or deleting snapshots, result in
      metadata updates to the file system using locks, and thus result in SCSI
      reservations.
      Reservations are also generated when you expand a virtual disk, when a
      snapshot for a virtual machine disk increases in size, or when a thin
      provisioned virtual disk grows.

      VMotion also use SCSI Reservations, with a reservation placed first by the
      source ESX, which is subsequently released, and then the destination ESX
      places a SCSI Reservation on the LUN.
      Note that SCSI reservations are used to acquire or release a lock on a file but
      is not required when updating host heartbeats to maintain a lock on a file.


     For more information, see KB 1005009: http://kb.vmware.com/kb/1005009



43                                       Confidential
What Should I do when I have SCSI Reservation Conflicts?

      Sometimes SCSI Reservation Conflicts are temporary and my subside. If the
      SCSI Reservation Conflicts do not subside, issue a LUN “RESET” to the LUN
      to see if this resolves them. This can be done from an ESX host by issuing the
      following command:
     # vmkfstools –L lunreset /vmfs/devices/disks/naa.xxxxxxxxxxxxxxxx

     *NOTE: To be safe, issue the LUN reset command at least twice.
      If the command was successful, you should see messages in the vmkernel
      similar to the following:
     cpu1:1057)<6>scsi(2:0:3:73): DEVICE RESET SUCCEEDED.

      If a LUN reset resolves the issue then this points to a lost reservation as
      being the root cause of the problem. A lost reservation means that an initiator
      placed the reservation however it was not able to send the release (HBA goes
      into fatal error state) or the release was dropped at some level yet a SUCCESS
      was returned to the initiator (blade environment).



44                                        Confidential
What Should I do when I have SCSI Reservation Conflicts?

      If the LUN reset does not resolve the SCSI reservation conflict, the
      following scenarios may apply:
       • Array is overloaded:
         • Write pending at 100%
         • Read/write cache exhausted
         • Cache constantly destaging down to disk
         • Array controller/port queue is full
         • LUN Replication/Backups too demanding
       • Host mode setting incorrect for initiator record on the array
       • Array software/firmware bug (LUN flag setting, Storage pool leak, etc)
       • Hardware Management Agents (HP Insight Manager) are locking the LUN
       • The LUN reporting the SCSI reservation is presented to a non-ESX host
       • The LUN reporting the SCSI reservation is an RDM LUN used by a VM


     *Remember: A SCSI reservation conflict is a symptom, not the problem.
45                                               Confidential
Best Practices to avoid SCSI Reservation Conflicts

      Keep HBA, Blade Switch (if applicable), Fabric Switch, and
      Array firmware up to date. Firmware updates contain fixes!
      Determine if any operations are already happening on the
      LUN you wish to start another operation that may cause a
      SCSI Reservation.
      Spread out VCB Backups, VMotion operations, Template
      deployments, etc.
      Choose one ESX Server as your deployment server to avoid
      conflicts with multiple ESX servers trying to deploy
      templates.
      Limit access Administrative operations in vCenter so that
      you control who can enact an operation that could lead to
      potential reservations issues (snapshots).



46                              Confidential
Best Practices to avoid SCSI Reservation Conflicts

      Avoid boot storms by scheduling VM reboots so that there
      is only one reboot per LUN. In the case of VMware View
      desktops, choose to leave these desktops always powered
      on.
      Use care when scheduling backups, antivirus, or LUN
      replication.
      Ensure that the storage array will not be overloaded with
      operations that could impact hosts connected.
      Verify that there is no other SCSI Reservations operation
      happening.
      Use the ATS (Atomic Test and Set) feature of VAAI, if
      available. More on this feature on the next slide.




47                               Confidential
Goodbye SCSI Reservations, Hello ATS!

 In vSphere 4.1 the following three offload primitives are supported
     – Hardware Assisted Locking, Full Copy, and Block Zero. These
     primitives are also known as Atomic Test and Set, Clone Blocks,
     and Write Same respectively. We will talk about Clone Blocks and
     Write Same in later slides.
 Atomic Test and Set (ATS) primitive atomically modifies a sector
     on disk without the use of SCSI reservations and needing to lock
     out other hosts from concurrent LUN access. This primitive also
     reduces the number of commands required to successfully acquire
     an on-disk lock.
 Upon receiving an ATS command, the array should atomically
     check if the contents of the disk block at the specified logical block
     number are the same as initiator-provided existing value, and if yes,
     replace it with initiator-provided new value.




48                                  Confidential
Goodbye SCSI Reservations, Hello ATS!

 Unlike SCSI reservations, ATS commands from other hosts should
     not be rejected with a device status of RESERVATION CONFLICT in
     the case of in-flight ATS commands from other hosts. Instead,
     these commands are queued as needed and processed in the same
     order.
 The ATS primitive can significantly improve I/O throughput from
     guest OS applications to a VMFS volume in the presence of
     metadata operations, and the number of concurrent cluster-aware
     VM operations that can be supported in the presence of these I/O
     intensive enterprise workloads.
 ATS will be used for the following operations:
     • Every lock operation executed by the VMFS-3 lock manager (metadata)
     • Power state operations (power on/off/checkpoint/resume, snapshot and
      consolidate) of VMs
     • Cluster-wide operations like Storage vMotion, vMotion, DRS


49                                      Confidential
Requirements for VAAI

 The following are required for VAAI:
     • ESX/ESXi 4.1
     • Array with SCSI ANSI version 3 or higher
     • Array firmware that supports VAAI primitives
 VAAI uses the following SCSI commands:
     • ATS uses SCSI command 0x89 (COMPARE AND WRITE)
     • Clone Blocks/Full Copy uses SCSI command 0x83 (EXTENDED COPY)
     • Zero Blocks/Write Same uses SCSI command 0x93 (WRITE SAME)
 These VAAI operations are controlled by the following advanced
     settings:
     • /VMFS3/HardwareAcceleratedLocking
     • /DataMover/HardwareAcceleratedMove
     • /DataMover/HardwareAcceleratedInit


50                                       Confidential
Known Issues for VAAI

 VAAI is enabled by default, and the primitives will issue the
     specific SCSI commands to determine whether the LUN supports
     the commands or not. If the commands are not successful, the LUN
     will be marked as non-VAAI capable.
 The reason VAAI is enabled by default is due to NDU (non-
     disruptive upgrades) or initiator setting changes that could make
     the LUNs support VAAI.
 This default behavior has caused issues for some arrays or fabric
     configurations that do not support the SCSI commands VAAI uses.
     Examples of this are the following:
     • Outdated/unsupported array firmware does not handle SCSI commands
      correctly and returns unexpected responses
     • Cisco SANTAP controllers do not support the commands and crash




51                                    Confidential
Known Issues for VAAI

 To check VAAI status, run the command:

     # esxcfg-scsidevs -l | egrep "Display Name:|VAAI Status:“
 In the event that your array or fabric configuration does not
     support VAAI commands and is behaving oddly, disable the VAAI
     functions on the ESX hosts until a firmware update is available
     from the vendor.
 To disable the VAAI primitives, execute the following commands
     on each ESX host:
     # esxcfg-advcfg -s 0 /DataMover/HardwareAcceleratedMove
     # esxcfg-advcfg -s 0 /DataMover/HardwareAcceleratedInit
     # esxcfg-advcfg -s 0 /VMFS3/HardwareAcceleratedLocking




52                                Confidential
Agenda
        vCenter Performance Charts & ESXTop
        Disk Alignment
        Hard Disks/Storage Array
        SCSI Reservations
        APD (All Paths Down) Issue
        Multipathing Considerations
        Communication




53                                  Confidential
APD (All Paths Down) Issue




54              Confidential
What is the APD issue?

      APD (all paths down) is a condition where all paths to the storage
      device (LUN) are lost either temporarily or permanently.
      The most common cause for the APD condition is due to
      intentional changes in the SAN configuration, particularly
      removing the presentation of a LUN to a give host(s) or migrating
      off a SAN entirely.
      The APD state can be transient, but in most cases it will be
      permanent, unless the removed LUN is presented back to the
      affected host, an advanced configuration option is used, or the
      host is rebooted to recover from the condition.
      The advanced configuration option was introduced in a patch after
      ESX 4.0 U1 was released (included in ESX 4.0 U2 and ESX 4.1):
      http://kb.vmware.com/kb/1016291
      The advanced configuration option can be enabled with the
      following command:
      # esxcfg-advcfg –s 1 /VMFS3/FailVolumeOpenIfAPD
55                                  Confidential
What is the impact of the APD issue?

      If a LUN has been removed from the ESX hosts improperly, this
      will lead to an APD state. While the LUN is in this state, running
      VM‟s may hang or pause momentarily due to the VMM requiring an
      fsLock. Issuing a rescan at this point will exasperate the issue and
      cause more VM‟s to pause and for a longer period of time.
      This hang/pause effect can cause severe issues for time
      sensitive/transaction applications such as SAP, SQL, or MS
      Exchange. It also effectively pauses the network stack of Windows
      based OS‟es and can cause ICMP ping drops as a result.
      This issue is discussed in detail via the following KB link:

      http://kb.vmware.com/kb/1016626




56                                   Confidential
How can I avoid the APD issue?

      Engineering has approved an official procedure for proper LUN
      removal in ESX 4.x.
      This process involves creating LUN claim rules that will mask the
      LUNs from the vmkernel before they are removed from the host.
      After masking the LUN, a reclaim command is run which will
      unregister those LUNs from the vmkernel, allowing them to be
      safely removed.
      Once the LUNs are removed, the claim rules that were added for
      this procedure should be removed.
      The proper LUN removal procedure to avoid the APD is fully
      covered in the following KB article:

      http://kb.vmware.com/kb/1015084




57                                  Confidential
Agenda
        vCenter Performance Charts & ESXTop
        Disk Alignment
        Hard Disks/Storage Array
        SCSI Reservations
        APD (All Paths Down) Issue
        Multipathing Considerations
        Communication




58                                  Confidential
Multipathing Considerations




59               Confidential
MRU (Most Recently Used) Path Selection Policy

 The MRU Path Selection Policy was designed for use with
     Active/Passive or Passive Not Ready (PNR) arrays. These arrays
     have a sense of LUN ownership, which means a LUN can only be
     accessed via one controller, not both. As such, Active/Passive
     arrays have far more failover conditions.
 The MRU policy selects the first working path discovered at system
     boot time. This would be the first HBA and first path discovered.
     This design has an inherit flaw as it means that unless failover has
     occurred, all I/O to the LUNs will be going through the first HBA.
 If this path becomes unavailable, the ESX host switches to an
     alternative path and continues to use the new path while it is
     available. It does this because forcing the path back to the other
     array controller would cause the LUN to change ownership on the
     array.




60                                 Confidential
Fixed Path Selection Policy

 Originally, the Fixed Path Selection Policy was primarily used by
     Active/Active arrays however this policy is now used by most
     ALUA compatible arrays as well as iSCSI arrays with a virtual port.
 The Fixed Path Policy uses the concept of a preferred path, which
     means that you can select which path you want to use for a LUN
     and that path will always be used unless a failover scenario occurs.
     Once the path is restored, however, the storage stack will
     automatically switch back to the preferred path.
 Fixed was used in the ESX 2.x – 3.x days to load balance I/O from
     an ESX perspective.
 It is acceptable to use the MRU policy for Active/Active arrays
     since both controllers can serve up a LUN, however it is NOT
     acceptable or supported to use Fixed policy on an Active/Passive
     or Passive Not Ready (PNR) array as this will cause path thrashing
     to occur.


61                                 Confidential
RoundRobin Path Selection Policy

 The RoundRobin Path Selection Policy is the only in-box policy
     that load balances I/O (Fixed doesn‟t count).
 This policy will only use “Active”, “Optimized” paths by default as
     valid paths to switch to. No “Standby” paths will be used as this
     would cause a trespassing effect/LUN ownership change on
     Active/Passive or Passive Not Ready (PNR) arrays to occur. This is
     not a problem for Active/Active arrays since they have no
     “Standby” paths.
 The load balancing mechanism will switch paths by one of two
     methods:
     • IOPS threshold (commands sent)
     • throughput threshold (bytes transferred).




62                                       Confidential
RoundRobin Path Selection Policy

 When setting the RoundRobin policy, the default value for IOPS will be
     1000 commands and the default throughput threshold will be 10485760
     bytes.
 VMware does not recommend settings for either the IOPS or the bytes
     value with Roundrobin. For information on the correct settings for
     RoundRobin use with a particular storage array, the storage array
     vendor MUST be contacted.
 The array vendor will know what settings are optimal for use but they
     will also know what settings NOT to use.
 Some vendors recommend an „iops‟ value of „1‟ to be used however
     this same setting on another array can have detrimental effects
     resulting in poor performance or even a possible production outage.
 The RoundRobin policy cannot be used on MSCS Quorum RDM LUNs.



63                                  Confidential
Vendor Recommend Round Robin Settings

 EMC DMX:
     http://www.emc.com/collateral/hardware/white-papers/h6531-using-
     vmware-vsphere-with-emc-symmetrix-wp.pdf
 EMC Celerra:
     http://www.emc.com/collateral/software/technical-
     documentation/h5536-vmware-esx-srvr-using-emc-celerra-stor-sys-
     wp.pdf
 NetApp:
     http://kb.vmware.com/kb/1010713
     http://media.netapp.com/documents/tr-3749.pdf
 HP EVA:
     http://h20195.www2.hp.com/v2/GetPDF.aspx/4AA1-2185ENW.pdf




64                               Confidential
Setting Round Robin Settings From CLI

 Setting an entire SATP (Storage Array Type Plugin) to Round
     Robin:
     # esxcli nmp satp setdefaultpsp -s VMW_SATP_LSI -p
     VMW_PSP_RR
 Setting the Round Robin PSP setting for a single LUN:
     # esxcli nmp device setpolicy -d
     naa.600174d001000000010f003048318438 --psp VMW_PSP_RR
 Command line option to set iops value:
     # esxcli nmp roundrobin setconfig -d
     naa.600174d001000000010f003048318438 --iops 10 --type
     iops
 Command line option to set bytes value:
     # esxcli nmp roundrobin setconfig -d
     naa.600174d001000000010f003048318438 --bytes 11 --
     type bytes


65                             Confidential
Displaying Round Robin Settings From CLI

 To displaying Round Robin settings for a LUN:
# esxcli nmp device naa.60a9800050334b356b4a51312f417541

       Device Display Name: NETAPP Fibre Channel Disk
     (naa.60a9800050334b356b4a51312f417541)

       Storage Array Type: VMW_SATP_ALUA

       Storage Array Type Device Config:

{implicit_support=on;explicit_support=off;explicit_allow=on;alua_fol
     lowover=on;{TPG_id=2,TPG_state=AO}{TPG_id=3,TPG_state=ANO}}

       Path Selection Policy: VMW_PSP_RR

       Path Selection Policy Device Config:
     {policy=rr,iops=1000,bytes=10485760,useANO=0;lastPathIndex=3:
     NumIOsPending=0,numBytesPending=0}

       Working Paths: vmhba2:C0:T2:L1, vmhba1:C0:T2:L1



66                                    Confidential
Known Issues for Round Robin in ESX 4.x

 There are currently two known issues when using the Round Robin
     Path Selection Policy:
     • After setting the IOPS value for roundrobin, the setting is not retained after a
      reboot and instead shows a much higher value. This is resolved in a patch for
      ESX 4.0:
      http://kb.vmware.com/kb/1017721
     • The SATP for EMC DMX (VMW_SATP_SYMM) does not distribute I/O evenly
      with roundrobin after a state change occurs. This state change includes adding
      a new path (FA) or when losing a paths and recovering. This is also resolved
      in a patch for ESX 4.0 and ESX 4.1:

      ESX 4.0:
      http://kb.vmware.com/kb/1023759

      ESX 4.1:
      http://kb.vmware.com/kb/1027013



67                                         Confidential
Agenda
        vCenter Performance Charts & ESXTop
        Disk Alignment
        Hard Disks/Storage Array
        SCSI Reservations
        APD (All Paths Down) Issue
        Multipathing Considerations
        Communication




68                                  Confidential
Communication




69        Confidential
Communication

 Communication between various business units is required for the
     success of a company.
 Many companies silo their teams into networking, storage,
     server/hardware, Windows/Linux administrators,
     Database/Exchange administrators, virtualization administrators,
     and so on.
 This silo design or syndrome has been responsible for lack of
     cooperation and communication between the siloed teams. There
     have been movements to try and break down these barriers
     however this is not an easy task.
 Ultimately, the victims of this design may not be the people of the
     silo teams, but the business/company itself, which in turn affects
     customers directly.




70                                 Confidential
Communication

 In recent years, virtualization has become a huge movement in
     datacenters and the criticality of virtualized servers have shifted
     from simple application servers to mission critical applications.
     This mission critical requirement is far more demanding from an
     I/O (CPU, memory, and disk) and disk capacity perspective.
 The need to have virtualization administrators and storage
     administrators communicate effectively has come to critical mass.
     More than ever, these two teams need to be on the same page at all
     times. The need to cross-pollinate knowledge between these teams
     has never been higher.
 Lack of effective communication only between these two silo
     groups has led some companies to experience production outages
     and scalability issues.




71                                  Confidential
Improving Communication

 While change is not easy, there are some simple policies that can
     be adopted in order to break down the silo walls between the
     storage and virtualization administration teams:
     • Cross-train between both teams (basic to intermediate)
     • Meetings where topics that could affect the other silo should always have at
      least one representative from that silo that understands the discussion
     • Regular meetings (weekly) between the two silos to discuss current and future
      projects
     • Email distribution list that contains all parties of both silos
     • Always consider the other party before any changes are made to the
      virtualization infrastructure, array configuration, or policies on either team




72                                           Confidential
Questions?




73           Confidential

Más contenido relacionado

La actualidad más candente

Design decision nfs-versus_fc_storage v_0.3
Design decision nfs-versus_fc_storage v_0.3Design decision nfs-versus_fc_storage v_0.3
Design decision nfs-versus_fc_storage v_0.3David Pasek
 
Avoiding Chaos: Methodology for Managing Performance in a Shared Storage A...
Avoiding Chaos:  Methodology for Managing Performance in a Shared Storage A...Avoiding Chaos:  Methodology for Managing Performance in a Shared Storage A...
Avoiding Chaos: Methodology for Managing Performance in a Shared Storage A...brettallison
 
Ceph Day Melbourne - Troubleshooting Ceph
Ceph Day Melbourne - Troubleshooting Ceph Ceph Day Melbourne - Troubleshooting Ceph
Ceph Day Melbourne - Troubleshooting Ceph Ceph Community
 
Revisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS SchedulerRevisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS SchedulerYongseok Oh
 
Introduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3RIntroduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3RSimon Huang
 
Moving to PCI Express based SSD with NVM Express
Moving to PCI Express based SSD with NVM ExpressMoving to PCI Express based SSD with NVM Express
Moving to PCI Express based SSD with NVM ExpressOdinot Stanislas
 
Ceph Day Beijing - SPDK for Ceph
Ceph Day Beijing - SPDK for CephCeph Day Beijing - SPDK for Ceph
Ceph Day Beijing - SPDK for CephDanielle Womboldt
 
IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...
IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...
IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...In-Memory Computing Summit
 
Impact of Intel Optane Technology on HPC
Impact of Intel Optane Technology on HPCImpact of Intel Optane Technology on HPC
Impact of Intel Optane Technology on HPCMemVerge
 
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...Patrick McGarry
 
The Magic of Hot Streaming Replication, Bruce Momjian
The Magic of Hot Streaming Replication, Bruce MomjianThe Magic of Hot Streaming Replication, Bruce Momjian
The Magic of Hot Streaming Replication, Bruce MomjianFuenteovejuna
 
XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...
XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...
XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...The Linux Foundation
 
Tuning DB2 in a Solaris Environment
Tuning DB2 in a Solaris EnvironmentTuning DB2 in a Solaris Environment
Tuning DB2 in a Solaris EnvironmentJignesh Shah
 
Varrow madness 2013 virtualizing sql presentation
Varrow madness 2013 virtualizing sql presentationVarrow madness 2013 virtualizing sql presentation
Varrow madness 2013 virtualizing sql presentationpittmantony
 
Conceitos de Capacity Planning e Sysplex por Fernando Ferreira
Conceitos de Capacity Planning e Sysplex por Fernando FerreiraConceitos de Capacity Planning e Sysplex por Fernando Ferreira
Conceitos de Capacity Planning e Sysplex por Fernando FerreiraJoao Galdino Mello de Souza
 
Vm13 vnx mixed workloads
Vm13 vnx mixed workloadsVm13 vnx mixed workloads
Vm13 vnx mixed workloadspittmantony
 
LizardFS-WhitePaper-Eng-v3.9.2-web
LizardFS-WhitePaper-Eng-v3.9.2-webLizardFS-WhitePaper-Eng-v3.9.2-web
LizardFS-WhitePaper-Eng-v3.9.2-webSzymon Haly
 
SOUG_SDM_OracleDB_V3
SOUG_SDM_OracleDB_V3SOUG_SDM_OracleDB_V3
SOUG_SDM_OracleDB_V3UniFabric
 

La actualidad más candente (20)

Design decision nfs-versus_fc_storage v_0.3
Design decision nfs-versus_fc_storage v_0.3Design decision nfs-versus_fc_storage v_0.3
Design decision nfs-versus_fc_storage v_0.3
 
Avoiding Chaos: Methodology for Managing Performance in a Shared Storage A...
Avoiding Chaos:  Methodology for Managing Performance in a Shared Storage A...Avoiding Chaos:  Methodology for Managing Performance in a Shared Storage A...
Avoiding Chaos: Methodology for Managing Performance in a Shared Storage A...
 
Ceph Day Melbourne - Troubleshooting Ceph
Ceph Day Melbourne - Troubleshooting Ceph Ceph Day Melbourne - Troubleshooting Ceph
Ceph Day Melbourne - Troubleshooting Ceph
 
SCSI-3 PGR Support on Symm
SCSI-3 PGR Support on SymmSCSI-3 PGR Support on Symm
SCSI-3 PGR Support on Symm
 
Revisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS SchedulerRevisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS Scheduler
 
Introduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3RIntroduction to NVMe Over Fabrics-V3R
Introduction to NVMe Over Fabrics-V3R
 
Moving to PCI Express based SSD with NVM Express
Moving to PCI Express based SSD with NVM ExpressMoving to PCI Express based SSD with NVM Express
Moving to PCI Express based SSD with NVM Express
 
Ceph Day Beijing - SPDK for Ceph
Ceph Day Beijing - SPDK for CephCeph Day Beijing - SPDK for Ceph
Ceph Day Beijing - SPDK for Ceph
 
IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...
IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...
IMCSummit 2015 - Day 1 Developer Session - The Science and Engineering Behind...
 
Impact of Intel Optane Technology on HPC
Impact of Intel Optane Technology on HPCImpact of Intel Optane Technology on HPC
Impact of Intel Optane Technology on HPC
 
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
Using Recently Published Ceph Reference Architectures to Select Your Ceph Con...
 
The Magic of Hot Streaming Replication, Bruce Momjian
The Magic of Hot Streaming Replication, Bruce MomjianThe Magic of Hot Streaming Replication, Bruce Momjian
The Magic of Hot Streaming Replication, Bruce Momjian
 
XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...
XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...
XPDS16: Live Migration of vGPU - Xiao Zheng, Intel Asia-Pacific Research & De...
 
Tuning DB2 in a Solaris Environment
Tuning DB2 in a Solaris EnvironmentTuning DB2 in a Solaris Environment
Tuning DB2 in a Solaris Environment
 
Varrow madness 2013 virtualizing sql presentation
Varrow madness 2013 virtualizing sql presentationVarrow madness 2013 virtualizing sql presentation
Varrow madness 2013 virtualizing sql presentation
 
Conceitos de Capacity Planning e Sysplex por Fernando Ferreira
Conceitos de Capacity Planning e Sysplex por Fernando FerreiraConceitos de Capacity Planning e Sysplex por Fernando Ferreira
Conceitos de Capacity Planning e Sysplex por Fernando Ferreira
 
Vm13 vnx mixed workloads
Vm13 vnx mixed workloadsVm13 vnx mixed workloads
Vm13 vnx mixed workloads
 
FAST VP Deep Dive
FAST VP Deep DiveFAST VP Deep Dive
FAST VP Deep Dive
 
LizardFS-WhitePaper-Eng-v3.9.2-web
LizardFS-WhitePaper-Eng-v3.9.2-webLizardFS-WhitePaper-Eng-v3.9.2-web
LizardFS-WhitePaper-Eng-v3.9.2-web
 
SOUG_SDM_OracleDB_V3
SOUG_SDM_OracleDB_V3SOUG_SDM_OracleDB_V3
SOUG_SDM_OracleDB_V3
 

Similar a Vstoragetamsupportday1 110311121032-phpapp02

vSphere vStorage: Troubleshooting Performance
vSphere vStorage: Troubleshooting PerformancevSphere vStorage: Troubleshooting Performance
vSphere vStorage: Troubleshooting PerformanceProfessionalVMware
 
Virtualization with Lenovo X6 Blade Servers: white paper
Virtualization with Lenovo X6 Blade Servers: white paperVirtualization with Lenovo X6 Blade Servers: white paper
Virtualization with Lenovo X6 Blade Servers: white paperLenovo Data Center
 
San 101 basics of administrating a san
San 101 basics of administrating a sanSan 101 basics of administrating a san
San 101 basics of administrating a sanpineapplebed24
 
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSAccelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSCeph Community
 
Presentation v mware v-sphere advanced troubleshooting by eric sloof
Presentation   v mware v-sphere advanced troubleshooting by eric sloofPresentation   v mware v-sphere advanced troubleshooting by eric sloof
Presentation v mware v-sphere advanced troubleshooting by eric sloofsolarisyourep
 
VMware Performance Troubleshooting
VMware Performance TroubleshootingVMware Performance Troubleshooting
VMware Performance Troubleshootingglbsolutions
 
Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...
Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...
Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...Виталий Стародубцев
 
SQLIO - measuring storage performance
SQLIO - measuring storage performanceSQLIO - measuring storage performance
SQLIO - measuring storage performancevalerian_ceaus
 
Steps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issuesSteps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issuesAshwin Pawar
 
Presentation v mware performance overview
Presentation   v mware performance overviewPresentation   v mware performance overview
Presentation v mware performance overviewsolarisyourep
 
SQLintersection keynote a tale of two teams
SQLintersection keynote a tale of two teamsSQLintersection keynote a tale of two teams
SQLintersection keynote a tale of two teamsSumeet Bansal
 
Quantifying Container Runtime Performance: OSCON 2017 Open Container Day
Quantifying Container Runtime Performance: OSCON 2017 Open Container DayQuantifying Container Runtime Performance: OSCON 2017 Open Container Day
Quantifying Container Runtime Performance: OSCON 2017 Open Container DayPhil Estes
 
JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021Gene Leyzarovich
 
USING EMC FAST SUITE WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMS
USING EMC FAST SUITE  WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMSUSING EMC FAST SUITE  WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMS
USING EMC FAST SUITE WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMSsuri1155
 
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Odinot Stanislas
 
Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...
Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...
Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...xKinAnx
 

Similar a Vstoragetamsupportday1 110311121032-phpapp02 (20)

vSphere vStorage: Troubleshooting Performance
vSphere vStorage: Troubleshooting PerformancevSphere vStorage: Troubleshooting Performance
vSphere vStorage: Troubleshooting Performance
 
Virtualization with Lenovo X6 Blade Servers: white paper
Virtualization with Lenovo X6 Blade Servers: white paperVirtualization with Lenovo X6 Blade Servers: white paper
Virtualization with Lenovo X6 Blade Servers: white paper
 
San 101 basics of administrating a san
San 101 basics of administrating a sanSan 101 basics of administrating a san
San 101 basics of administrating a san
 
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSAccelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDS
 
3487570
34875703487570
3487570
 
Presentation v mware v-sphere advanced troubleshooting by eric sloof
Presentation   v mware v-sphere advanced troubleshooting by eric sloofPresentation   v mware v-sphere advanced troubleshooting by eric sloof
Presentation v mware v-sphere advanced troubleshooting by eric sloof
 
NVMe over Fabric
NVMe over FabricNVMe over Fabric
NVMe over Fabric
 
os
osos
os
 
VMware Performance Troubleshooting
VMware Performance TroubleshootingVMware Performance Troubleshooting
VMware Performance Troubleshooting
 
Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...
Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...
Технологии работы с дисковыми хранилищами и файловыми системами Windows Serve...
 
SQLIO - measuring storage performance
SQLIO - measuring storage performanceSQLIO - measuring storage performance
SQLIO - measuring storage performance
 
Steps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issuesSteps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issues
 
Presentation v mware performance overview
Presentation   v mware performance overviewPresentation   v mware performance overview
Presentation v mware performance overview
 
SQLintersection keynote a tale of two teams
SQLintersection keynote a tale of two teamsSQLintersection keynote a tale of two teams
SQLintersection keynote a tale of two teams
 
Quantifying Container Runtime Performance: OSCON 2017 Open Container Day
Quantifying Container Runtime Performance: OSCON 2017 Open Container DayQuantifying Container Runtime Performance: OSCON 2017 Open Container Day
Quantifying Container Runtime Performance: OSCON 2017 Open Container Day
 
JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021JetStor portfolio update final_2020-2021
JetStor portfolio update final_2020-2021
 
USING EMC FAST SUITE WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMS
USING EMC FAST SUITE  WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMSUSING EMC FAST SUITE  WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMS
USING EMC FAST SUITE WITH SYBASE ASE ON EMC VNX STORAGE SYSTEMS
 
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
Hands-on Lab: How to Unleash Your Storage Performance by Using NVM Express™ B...
 
Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...
Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...
Ibm spectrum scale fundamentals workshop for americas part 7 spectrumscale el...
 
2011 q1-indy-vmug
2011 q1-indy-vmug2011 q1-indy-vmug
2011 q1-indy-vmug
 

Más de Suresh Kumar

Vsphere 4-partner-training180
Vsphere 4-partner-training180Vsphere 4-partner-training180
Vsphere 4-partner-training180Suresh Kumar
 
Vsphere4 100325065654-phpapp01
Vsphere4 100325065654-phpapp01Vsphere4 100325065654-phpapp01
Vsphere4 100325065654-phpapp01Suresh Kumar
 
Vmwareserver tips-tricks-110218231744-phpapp01
Vmwareserver tips-tricks-110218231744-phpapp01Vmwareserver tips-tricks-110218231744-phpapp01
Vmwareserver tips-tricks-110218231744-phpapp01Suresh Kumar
 
Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)
Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)
Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)Suresh Kumar
 
Vmwareperformancetroubleshooting 100224104321-phpapp02
Vmwareperformancetroubleshooting 100224104321-phpapp02Vmwareperformancetroubleshooting 100224104321-phpapp02
Vmwareperformancetroubleshooting 100224104321-phpapp02Suresh Kumar
 
Managingvspherewiththevesi 091210144626-phpapp02
Managingvspherewiththevesi 091210144626-phpapp02Managingvspherewiththevesi 091210144626-phpapp02
Managingvspherewiththevesi 091210144626-phpapp02Suresh Kumar
 
Advancedtroubleshooting 101208145718-phpapp01
Advancedtroubleshooting 101208145718-phpapp01Advancedtroubleshooting 101208145718-phpapp01
Advancedtroubleshooting 101208145718-phpapp01Suresh Kumar
 
Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02
Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02
Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02Suresh Kumar
 

Más de Suresh Kumar (8)

Vsphere 4-partner-training180
Vsphere 4-partner-training180Vsphere 4-partner-training180
Vsphere 4-partner-training180
 
Vsphere4 100325065654-phpapp01
Vsphere4 100325065654-phpapp01Vsphere4 100325065654-phpapp01
Vsphere4 100325065654-phpapp01
 
Vmwareserver tips-tricks-110218231744-phpapp01
Vmwareserver tips-tricks-110218231744-phpapp01Vmwareserver tips-tricks-110218231744-phpapp01
Vmwareserver tips-tricks-110218231744-phpapp01
 
Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)
Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)
Vmwareperformancetroubleshooting 100224104321-phpapp02 (1)
 
Vmwareperformancetroubleshooting 100224104321-phpapp02
Vmwareperformancetroubleshooting 100224104321-phpapp02Vmwareperformancetroubleshooting 100224104321-phpapp02
Vmwareperformancetroubleshooting 100224104321-phpapp02
 
Managingvspherewiththevesi 091210144626-phpapp02
Managingvspherewiththevesi 091210144626-phpapp02Managingvspherewiththevesi 091210144626-phpapp02
Managingvspherewiththevesi 091210144626-phpapp02
 
Advancedtroubleshooting 101208145718-phpapp01
Advancedtroubleshooting 101208145718-phpapp01Advancedtroubleshooting 101208145718-phpapp01
Advancedtroubleshooting 101208145718-phpapp01
 
Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02
Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02
Advancedperformancetroubleshootingusingesxtop 101110131727-phpapp02
 

Último

Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...gurkirankumar98700
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure servicePooja Nehwal
 

Último (20)

Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
 

Vstoragetamsupportday1 110311121032-phpapp02

  • 1. vStorage: Troubleshooting Performance Nathan Small, Staff Engineer, Global Support Services Confidential © 2009 VMware Inc. All rights reserved
  • 2. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 2 Confidential
  • 3. vCenter Performance Charts & ESXTop 3 Confidential
  • 4. Performance Charts  Through the Overview section of the Performance tab of an ESX host, we can observe if a host is seeing recent disk latency issues.  Unfortunately, this chart does not pinpoint which device or path is seeing this latency. For a more granularity, we will use “Advanced” 4 Confidential
  • 5. Performance Charts  If we switch to the “Storage path” option, we can see latency is coming from vmhba0:C0:T0:L0 for both read and write latency: 5 Confidential
  • 6. Performance Charts  In order to see the latency statistics for a given LUN, we will need to go to the “Disk” view and then select “Chart Options”: Select the LUNs you wish to see the latency statistics for under “Objects”, then select the Physical device read and write latency “Counters” 6 Confidential
  • 7. Performance Charts  With this granular view, we can see that the LUN with ID naa.60060480000190300919533030303134 had a latency spike up to 48ms at ~9:50am 7 Confidential
  • 8. ESXTop  ESXTop is a great tool to use if you need performance captured during a specific timeframe (scripted batch mode) or if the ESX host is unavailable in vCenter/VI Client.  To view the equivalent of the “Storage path” view, hit „d‟: 8 Confidential
  • 9. ESXTop  You can toggle the display fields by hitting „f‟. By taking out unneeded fields (I/O stats), you can make the display screen far easier to read and work with: 9 Confidential
  • 10. ESXTop  To view individual LUN statistics (with unique ID), hit „u‟:  As we can see, the default field size for “DEVICE” is not long enough to show the entire unique identifier. To change this field size, hit „L‟:  10 Confidential
  • 11. ESXTop  Expanding this field now shows the entire unique ID for the LUN and NFS datastores:  Just as before, we should display only the fields we need to be concerned with by hitting „f‟: 11 Confidential
  • 12. ESXTop  This view now shows us similar output to the „d‟ view except we see the statistics for a LUN (all paths). *Note: The reason there are no statistics for the NFS mounts is due to the fact that these are ISO and template repositories that are not in use currently by this host. 12 Confidential
  • 13. ESXTop  To capture esxtop data, use „vm-support -S‟. Here is an example for collecting these statistics at 2 second intervals over 20 seconds: # vm-support –S –i 2 –d 20 13 Confidential
  • 14. ESXTop  To replay that capture data, extract the compressed file, change to the „snaphots‟ directory, run the uncompress script, then run „esxtop‟ in replay mode „-R‟: 14 Confidential
  • 15. DAVG/KAVG/GAVG/QAVG  What is the DAVG? • The DAVG field is the Device Average, which is the amount of time it takes for a SCSI command to leave the HBA, hit the array, and return completed. This is the most effective method of determining if the performance issue being experienced is at the physical layer (switch/array).  What are acceptable DAVG values? • Optimal, sustained DAVG latency values would be between 0-5ms for 4/8GB FC and 10GB iSCSI/NFS. Seeing 5-10ms latency for 1/2GB FC or 1GB iSCSI/NFS is also acceptable. • Seeing latency values of 10-20ms could start to show some minor performance issues inside the VM. while values of 20-50ms would show noticeable to significant performance issues. Latency values of 50ms or higher would make the VMs almost unusable.  How should I proceed when my DAVG is sustained at a high level? • Engage the storage team immediately. They may already be aware of the latency due to increased load on the storage array (LUN replication, backups) 15 Confidential
  • 16. DAVG/KAVG/GAVG/QAVG  What is the KAVG? • The KAVG field is the Kernel Average, which is the amount of time a SCSI command spends in the vmkernel.  What are acceptable KAVG values? • The KAVG should always be less than 1ms.  How should I proceed when my KAVG is sustained at a high level? • While rare to see the KAVG value high at all, it has been observed when the queue has been throttled back on the ESX host due to a TASK_SET_FULL condition on the array (KB 1008113), or the queue depth has been reduced on the HBAs due to configuration restrictions (KB 1006001). It has also been observed temporarily on path failover. This is due to commands queuing during the path activation process. 16 Confidential
  • 17. DAVG/KAVG/GAVG/QAVG  What is the GAVG? • The GAVG field is the Device Average + Kernel Average + Queue Average, which is the latency perceived by the Guest OS or VM.  What are acceptable GAVG values? • As the KAVG and QAVG should always be less than 1ms, acceptable GAVG latency values would be the same as acceptable DAVG values.  How should I proceed when my GAVG is sustained at a high level? • The same steps should be followed for high DAVG values. 17 Confidential
  • 18. DAVG/KAVG/GAVG/QAVG  What is the QAVG? • The QAVG field is the Queue Average, which is the amount of time the SCSI command spends in the HBA driver.  What are acceptable QAVG values? • The QAVG should always be less than 1ms.  How should I proceed when my QAVG is sustained at a high level? • Just like the KAVG, the QAVG should not have a high latency value, unless there is a legitimate reason for commands to remained queued on the host side. 18 Confidential
  • 19. Queue Depth and ESXTop  Many people believe that increasing the queue depth will solve a performance issue or improve performance overall, however this can have the opposite effect (KB 1006001 & 1008113).  The queue depth should only be increased if the array vendor recommends to do so or the queue depth is getting exhausted on the host side.  As seen in a previous slide, we can observe the queue depth usage and activity: 19 Confidential
  • 20. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 20 Confidential
  • 21. Disk Alignment 21 Confidential
  • 22. Disk Alignment  The concept of Disk Alignment is simple, yet the performance hit for misaligned I/O can be troublesome for a single, high disk I/O VM (SQL, Exchange, etc) or for the entire virtualized infrastructure on a particular storage array.  It has been a best practice recommendation from high I/O application vendors to ensure that the data drive for the application is aligned to ensure the best I/O possible, as this will avoid any partial read or write operation. In many cases, this data drive would be a separate LUN either on local storage or on a share storage array. 22 Confidential
  • 23. Disk Alignment  In a SAN environment, the smallest hardware unit used by a SAN storage array to build a LUN out of multiple physical disks is a called a chunk or a stripe. To optimize I/O, chunks are usually much larger than sectors. Thus a SCSI I/O request that intends to read a sector in reality reads one chunk.  On top of this, in a Windows environment NTFS is formatted in blocks ranging from 1MB to 8MB. The file system used by the guest operating system optimizes I/O by grouping sectors into so- called clusters (allocation units).  Older operating systems on x86 architectures create partitions with a master boot record (MBR) that consumes 63 sectors. This is due to legacy BIOS code from the PC that used cylinder, head, and sector addressing instead of logical block addressing (LBA). Without LBA, the first track is reserved for the boot code, and the first partition starts at cylinder 0, head 1, and sector 1. This is LBA 63 and is therefore unaligned. 23 Confidential
  • 24. Disk Alignment  An unaligned partition results in a track crossing and an additional I/O, incurring a penalty on latency and throughput. The additional I/O (especially if small) can impact system resources significantly on some host types. An aligned partition ensures that the single I/O is serviced by a single device, eliminating the additional I/O and resulting in overall performance improvement.  Unaligned structure may cause many additional I/O operations when only one cluster is ready by the guest operating system. 24 Confidential
  • 25. Disk Alignment  Beginning with ESX 3.0.x, or rather VMFS3, any newly created VMFS LUN will be aligned to a starting sector of 128, which is 64k (128 * 512 = 65536). As most arrays perform reads/writes in 4k, or a value divisible by 4k, this means that the VMFS partition is aligned with the underlying storage. When the partition for the VM is aligned in the same manner, I/O is aligned across the board.  I/O improvements on a properly aligned Windows NTFS volume in a VMDK on a SAN LUN: 25 Confidential
  • 26. Disk Alignment  The result of aligning I/O can have significant, measurable performance improvements for high disk I/O VMs:  VMW  SQL 26 Confidential
  • 27. Disk Alignment and NFS  Does Disk Alignment only apply to VMFS and not NFS?  While the NFS volume does not have an additional VMFS layer to align, the NFS datastore is aligned to a 4k divisible offset when created and the array controllers still operate in 4k read/write operations.  As such, there is still a need to align the OS and data drives of VMs residing on an NFS volume. 27 Confidential
  • 28. Verifying Disk Alignment  How can I tell if the disks inside my VM‟s are aligned?  This is the default partition offset of 63: 63 * 512 = 32256 28 Confidential
  • 29. Aligning Disks  To align newly created disk (data disks), simply use diskpart (Windows) or fdisk (Linux). For more information, see: http://www.vmware.com/pdf/esx3_partition_align.pdf http://media.netapp.com/documents/tr-3747.pdf  Aligning existing disks gets tricky since the file system and data has already been written. To simply change the offset of these existing partitions with diskpart or fdisk will render them useless.  vSphere/ESX currently has no way of aligning disks however 3rd party vendors have created utilities to do this: • Netapp provides a utility called ‘mbrscan’, which will scan your VMs’ disks to see if they are aligned, and ‘mbralign’, which will align the disks of a VM (powered off VMs only). • Vizioncore makes a product called vOptimizer that will align VM’s disks to 64k offset. Alternatively, their vConverter 5.0 product will convert physical or virtual machines and in the process align the disk offset to 64k • Paragon also makes a utility called Paragon Alignment Tool (PAT). This also comes bundled with their flagship product. 29 Confidential
  • 30. Planning Ahead  The best way to plan ahead with Disk Alignment is to create templates for every OS you wish to deploy and align those before installing the OS. This way you can deploy VM‟s that will already be aligned. 30 Confidential
  • 31. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 31 Confidential
  • 33. SCSI vs. SATA  The argument of which hard drive technology to use is a well known discussion however the answer on which drives to use ultimately depends on the application or need.  SCSI drives are more expensive than SATA drives however this is due to the design and performance achieved with this technology.  SCSI uses a tagged command queuing implementation which allows many commands to be outstanding. This design provides significant performance gain for the drives/controllers to be able to order these commands in most optimal execution manner possible.  SCSI drives also use a processor for executing commands and handling the interface while a separate processor handles the head positioning through servos.  SCSI can push a maximum of 640Meg/second 33 Confidential
  • 34. SCSI vs. SATA  In the early years, SATA used Task Command Queuing (TCQ). This queuing technology was designed to help bridge the gap between SCSI and PATA/SATA drives however there was too much overhead and TCQ simply wasn‟t efficient enough. TCQ is sometimes referred to as interrupt queuing.  SATA 2.0 introduced Native Command Queuing (NCQ), which is very similar to the outstanding request queuing that SCSI uses. This technology drastically reduces the amount of interrupts that TCQ uses due to the integrated first party DMA engine as well as the intelligent ordering of commands and interrupting when most it is most optimal to do so.  Due to NCQ and the throughput increases with the SATA 2.0 standard, SATA can perform well but not to the same level as SCSI.  SATA 3GB/s can push a maximum of 300MB/s 34 Confidential
  • 35. When To Use SCSI  SCSI drives are by far the most expensive choice, however you get what you pay for: Performance.  There are no limitations for SCSI drives usage however budgets are rarely unlimited so SCSI drives are usually purchased for production servers/applications (SQL, MS Exchange, etc) or tier 1/2 storage.  Because of the throughput capabilities of SCSI, these drives will handle I/O bursts far better than SATA, which would rely heavily on read/write cache to rough the storm.  SCSI drives are recommended for use with VMware View environments for linked clone and fully deployed desktops. More on this later. 35 Confidential
  • 36. When To Use SATA  While SATA drives are far cheaper than SCSI drives and usually higher in disk capacity, they simply do not perform at the same level as SCSI drives.  Installing more write cache in the storage array can assist SATA drives through heavy I/O periods however that would make write cache the last line of defence to an I/O storm or increasing I/O load over time. It may only be a matter of time before write cache fills up and then begins the battle of destaging.  If SATA disks must be used, a configuration of RAID 10 for those disks is strongly recommended. This will not guarantee availability in the heaviest of I/O storms, but it is the most optimal RAID configuration while maintaining redundancy. 36 Confidential
  • 37. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 37 Confidential
  • 38. SCSI Reservations 38 Confidential
  • 39. What are SCSI Reservations?  SCSI Reservations refer to the act of an initiator to using SCSI-2 commands 0x16 (RESERVE) and 0x17 (RELEASE) to lock a LUN for a specific operation.  While a reservation is placed on a LUN, if another initiator attempts to perform any command on the reserved LUN other than an INQUIRY, REQUEST SENSE, or a PREVENT/ALLOW MEDIUM REMOVAL command the command shall be rejected with RESERVATION CONFLICT device status.  A reservation may only be released by the initiator that placed the reservation. A release command sent by another initiator will be ignored.  Any initiator can clear the reservation by issuing a “BUS DEVICE RESET” command or a hard “RESET”. 39 Confidential
  • 40. SCSI Reservations  Since VMFS-3 is basically a clustered file system, allowing simultaneous access from multiple ESX servers, we need a way of preserving the integrity of the file system when more than one host is updating it.  VMFS-3 implements a locking protocol to prevent VMware cluster-aware applications from powering on (or otherwise sharing) the contents of a given virtual disk on more than one host at any given time.  Part of this locking protocol is based on the notion of on- disk locks that protect the metadata on the volume.  An ESX/ESXi host interested in locked access to metadata must atomically check the owner and lock-state fields in the lock, and if the lock is free, acquire it by writing out its own ID and lock state into the owner and lock-state fields. 40 Confidential
  • 41. SCSI Reservations  The lock manager currently uses SCSI reservations to check the lock information, establish itself as the owner by writing to the relevant lock fields on disk if the lock is free, and then releases the SCSI reservation.  ESX uses SCSI-2 non-persistent reservations which implies that only a single host can reserve the LUN in question at any one time. A reboot of that host will clear the reservation on the LUN (non-persistent).  ESX 4.0 introduces limited support for SCSI-3 persistent reservations to work with Microsoft Windows 2008 Clustering, however the LSISAS controller must be used for the VMs. Only Windows 2k8 is supported with this controller.  VMFS lock operations are still implemented using SCSI-2 non-persistent reservations, even in ESX 4.x. 41 Confidential
  • 42. A typical SCSI Reservation Conflict message vmkernel: 0:12:44:32.598 cpu4:1046)SCSI: vm 1046: 5509: Sync CR at 64 vmkernel: 0:12:44:33.520 cpu4:1046)SCSI: vm 1046: 5509: Sync CR at 48 vmkernel: 0:12:44:34.512 cpu4:1046)SCSI: vm 1046: 5509: Sync CR at 32 vmkernel: 0:12:44:35.482 cpu4:1046)SCSI: vm 1046: 5509: Sync CR at 16 vmkernel: 0:12:44:36.490 cpu2:1046)SCSI: vm 1046: 5509: Sync CR at 0 vmkernel: 0:12:44:36.490 cpu2:1046)WARNING: SCSI: 5519: Failing I/O due to too many reservation conflicts vmkernel: 0:12:44:36.490 cpu2:1046)WARNING: SCSI: 5615: status SCSI reservation conflict, rstatus 0xc0de01 for vmhba2:0:3. residual R 919, CR 0, ER 3 vmkernel: 0:12:44:36.490 cpu2:1046)FSS: 343: Failed with status 0xbad0022 for f530 28 2 45378334 5f8825b2 17007a7d 1d624ca4 4 1 0 0 0 0 0 "CR" stands for conflict retry. In this case an I/O is being retried due to reservation conflicts. The number at the end of the log statement is the number of retries left. In this case, all retries were exhausted so the I/O failed. 42 Confidential
  • 43. What causes a SCSI Reservation?  Administrative operations, such as creating or deleting a virtual disk, extending a VMFS volume, or creating or deleting snapshots, result in metadata updates to the file system using locks, and thus result in SCSI reservations.  Reservations are also generated when you expand a virtual disk, when a snapshot for a virtual machine disk increases in size, or when a thin provisioned virtual disk grows.  VMotion also use SCSI Reservations, with a reservation placed first by the source ESX, which is subsequently released, and then the destination ESX places a SCSI Reservation on the LUN.  Note that SCSI reservations are used to acquire or release a lock on a file but is not required when updating host heartbeats to maintain a lock on a file. For more information, see KB 1005009: http://kb.vmware.com/kb/1005009 43 Confidential
  • 44. What Should I do when I have SCSI Reservation Conflicts?  Sometimes SCSI Reservation Conflicts are temporary and my subside. If the SCSI Reservation Conflicts do not subside, issue a LUN “RESET” to the LUN to see if this resolves them. This can be done from an ESX host by issuing the following command: # vmkfstools –L lunreset /vmfs/devices/disks/naa.xxxxxxxxxxxxxxxx *NOTE: To be safe, issue the LUN reset command at least twice.  If the command was successful, you should see messages in the vmkernel similar to the following: cpu1:1057)<6>scsi(2:0:3:73): DEVICE RESET SUCCEEDED.  If a LUN reset resolves the issue then this points to a lost reservation as being the root cause of the problem. A lost reservation means that an initiator placed the reservation however it was not able to send the release (HBA goes into fatal error state) or the release was dropped at some level yet a SUCCESS was returned to the initiator (blade environment). 44 Confidential
  • 45. What Should I do when I have SCSI Reservation Conflicts?  If the LUN reset does not resolve the SCSI reservation conflict, the following scenarios may apply: • Array is overloaded: • Write pending at 100% • Read/write cache exhausted • Cache constantly destaging down to disk • Array controller/port queue is full • LUN Replication/Backups too demanding • Host mode setting incorrect for initiator record on the array • Array software/firmware bug (LUN flag setting, Storage pool leak, etc) • Hardware Management Agents (HP Insight Manager) are locking the LUN • The LUN reporting the SCSI reservation is presented to a non-ESX host • The LUN reporting the SCSI reservation is an RDM LUN used by a VM *Remember: A SCSI reservation conflict is a symptom, not the problem. 45 Confidential
  • 46. Best Practices to avoid SCSI Reservation Conflicts  Keep HBA, Blade Switch (if applicable), Fabric Switch, and Array firmware up to date. Firmware updates contain fixes!  Determine if any operations are already happening on the LUN you wish to start another operation that may cause a SCSI Reservation.  Spread out VCB Backups, VMotion operations, Template deployments, etc.  Choose one ESX Server as your deployment server to avoid conflicts with multiple ESX servers trying to deploy templates.  Limit access Administrative operations in vCenter so that you control who can enact an operation that could lead to potential reservations issues (snapshots). 46 Confidential
  • 47. Best Practices to avoid SCSI Reservation Conflicts  Avoid boot storms by scheduling VM reboots so that there is only one reboot per LUN. In the case of VMware View desktops, choose to leave these desktops always powered on.  Use care when scheduling backups, antivirus, or LUN replication.  Ensure that the storage array will not be overloaded with operations that could impact hosts connected.  Verify that there is no other SCSI Reservations operation happening.  Use the ATS (Atomic Test and Set) feature of VAAI, if available. More on this feature on the next slide. 47 Confidential
  • 48. Goodbye SCSI Reservations, Hello ATS!  In vSphere 4.1 the following three offload primitives are supported – Hardware Assisted Locking, Full Copy, and Block Zero. These primitives are also known as Atomic Test and Set, Clone Blocks, and Write Same respectively. We will talk about Clone Blocks and Write Same in later slides.  Atomic Test and Set (ATS) primitive atomically modifies a sector on disk without the use of SCSI reservations and needing to lock out other hosts from concurrent LUN access. This primitive also reduces the number of commands required to successfully acquire an on-disk lock.  Upon receiving an ATS command, the array should atomically check if the contents of the disk block at the specified logical block number are the same as initiator-provided existing value, and if yes, replace it with initiator-provided new value. 48 Confidential
  • 49. Goodbye SCSI Reservations, Hello ATS!  Unlike SCSI reservations, ATS commands from other hosts should not be rejected with a device status of RESERVATION CONFLICT in the case of in-flight ATS commands from other hosts. Instead, these commands are queued as needed and processed in the same order.  The ATS primitive can significantly improve I/O throughput from guest OS applications to a VMFS volume in the presence of metadata operations, and the number of concurrent cluster-aware VM operations that can be supported in the presence of these I/O intensive enterprise workloads.  ATS will be used for the following operations: • Every lock operation executed by the VMFS-3 lock manager (metadata) • Power state operations (power on/off/checkpoint/resume, snapshot and consolidate) of VMs • Cluster-wide operations like Storage vMotion, vMotion, DRS 49 Confidential
  • 50. Requirements for VAAI  The following are required for VAAI: • ESX/ESXi 4.1 • Array with SCSI ANSI version 3 or higher • Array firmware that supports VAAI primitives  VAAI uses the following SCSI commands: • ATS uses SCSI command 0x89 (COMPARE AND WRITE) • Clone Blocks/Full Copy uses SCSI command 0x83 (EXTENDED COPY) • Zero Blocks/Write Same uses SCSI command 0x93 (WRITE SAME)  These VAAI operations are controlled by the following advanced settings: • /VMFS3/HardwareAcceleratedLocking • /DataMover/HardwareAcceleratedMove • /DataMover/HardwareAcceleratedInit 50 Confidential
  • 51. Known Issues for VAAI  VAAI is enabled by default, and the primitives will issue the specific SCSI commands to determine whether the LUN supports the commands or not. If the commands are not successful, the LUN will be marked as non-VAAI capable.  The reason VAAI is enabled by default is due to NDU (non- disruptive upgrades) or initiator setting changes that could make the LUNs support VAAI.  This default behavior has caused issues for some arrays or fabric configurations that do not support the SCSI commands VAAI uses. Examples of this are the following: • Outdated/unsupported array firmware does not handle SCSI commands correctly and returns unexpected responses • Cisco SANTAP controllers do not support the commands and crash 51 Confidential
  • 52. Known Issues for VAAI  To check VAAI status, run the command: # esxcfg-scsidevs -l | egrep "Display Name:|VAAI Status:“  In the event that your array or fabric configuration does not support VAAI commands and is behaving oddly, disable the VAAI functions on the ESX hosts until a firmware update is available from the vendor.  To disable the VAAI primitives, execute the following commands on each ESX host: # esxcfg-advcfg -s 0 /DataMover/HardwareAcceleratedMove # esxcfg-advcfg -s 0 /DataMover/HardwareAcceleratedInit # esxcfg-advcfg -s 0 /VMFS3/HardwareAcceleratedLocking 52 Confidential
  • 53. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 53 Confidential
  • 54. APD (All Paths Down) Issue 54 Confidential
  • 55. What is the APD issue?  APD (all paths down) is a condition where all paths to the storage device (LUN) are lost either temporarily or permanently.  The most common cause for the APD condition is due to intentional changes in the SAN configuration, particularly removing the presentation of a LUN to a give host(s) or migrating off a SAN entirely.  The APD state can be transient, but in most cases it will be permanent, unless the removed LUN is presented back to the affected host, an advanced configuration option is used, or the host is rebooted to recover from the condition.  The advanced configuration option was introduced in a patch after ESX 4.0 U1 was released (included in ESX 4.0 U2 and ESX 4.1): http://kb.vmware.com/kb/1016291  The advanced configuration option can be enabled with the following command: # esxcfg-advcfg –s 1 /VMFS3/FailVolumeOpenIfAPD 55 Confidential
  • 56. What is the impact of the APD issue?  If a LUN has been removed from the ESX hosts improperly, this will lead to an APD state. While the LUN is in this state, running VM‟s may hang or pause momentarily due to the VMM requiring an fsLock. Issuing a rescan at this point will exasperate the issue and cause more VM‟s to pause and for a longer period of time.  This hang/pause effect can cause severe issues for time sensitive/transaction applications such as SAP, SQL, or MS Exchange. It also effectively pauses the network stack of Windows based OS‟es and can cause ICMP ping drops as a result.  This issue is discussed in detail via the following KB link: http://kb.vmware.com/kb/1016626 56 Confidential
  • 57. How can I avoid the APD issue?  Engineering has approved an official procedure for proper LUN removal in ESX 4.x.  This process involves creating LUN claim rules that will mask the LUNs from the vmkernel before they are removed from the host.  After masking the LUN, a reclaim command is run which will unregister those LUNs from the vmkernel, allowing them to be safely removed.  Once the LUNs are removed, the claim rules that were added for this procedure should be removed.  The proper LUN removal procedure to avoid the APD is fully covered in the following KB article: http://kb.vmware.com/kb/1015084 57 Confidential
  • 58. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 58 Confidential
  • 60. MRU (Most Recently Used) Path Selection Policy  The MRU Path Selection Policy was designed for use with Active/Passive or Passive Not Ready (PNR) arrays. These arrays have a sense of LUN ownership, which means a LUN can only be accessed via one controller, not both. As such, Active/Passive arrays have far more failover conditions.  The MRU policy selects the first working path discovered at system boot time. This would be the first HBA and first path discovered. This design has an inherit flaw as it means that unless failover has occurred, all I/O to the LUNs will be going through the first HBA.  If this path becomes unavailable, the ESX host switches to an alternative path and continues to use the new path while it is available. It does this because forcing the path back to the other array controller would cause the LUN to change ownership on the array. 60 Confidential
  • 61. Fixed Path Selection Policy  Originally, the Fixed Path Selection Policy was primarily used by Active/Active arrays however this policy is now used by most ALUA compatible arrays as well as iSCSI arrays with a virtual port.  The Fixed Path Policy uses the concept of a preferred path, which means that you can select which path you want to use for a LUN and that path will always be used unless a failover scenario occurs. Once the path is restored, however, the storage stack will automatically switch back to the preferred path.  Fixed was used in the ESX 2.x – 3.x days to load balance I/O from an ESX perspective.  It is acceptable to use the MRU policy for Active/Active arrays since both controllers can serve up a LUN, however it is NOT acceptable or supported to use Fixed policy on an Active/Passive or Passive Not Ready (PNR) array as this will cause path thrashing to occur. 61 Confidential
  • 62. RoundRobin Path Selection Policy  The RoundRobin Path Selection Policy is the only in-box policy that load balances I/O (Fixed doesn‟t count).  This policy will only use “Active”, “Optimized” paths by default as valid paths to switch to. No “Standby” paths will be used as this would cause a trespassing effect/LUN ownership change on Active/Passive or Passive Not Ready (PNR) arrays to occur. This is not a problem for Active/Active arrays since they have no “Standby” paths.  The load balancing mechanism will switch paths by one of two methods: • IOPS threshold (commands sent) • throughput threshold (bytes transferred). 62 Confidential
  • 63. RoundRobin Path Selection Policy  When setting the RoundRobin policy, the default value for IOPS will be 1000 commands and the default throughput threshold will be 10485760 bytes.  VMware does not recommend settings for either the IOPS or the bytes value with Roundrobin. For information on the correct settings for RoundRobin use with a particular storage array, the storage array vendor MUST be contacted.  The array vendor will know what settings are optimal for use but they will also know what settings NOT to use.  Some vendors recommend an „iops‟ value of „1‟ to be used however this same setting on another array can have detrimental effects resulting in poor performance or even a possible production outage.  The RoundRobin policy cannot be used on MSCS Quorum RDM LUNs. 63 Confidential
  • 64. Vendor Recommend Round Robin Settings  EMC DMX: http://www.emc.com/collateral/hardware/white-papers/h6531-using- vmware-vsphere-with-emc-symmetrix-wp.pdf  EMC Celerra: http://www.emc.com/collateral/software/technical- documentation/h5536-vmware-esx-srvr-using-emc-celerra-stor-sys- wp.pdf  NetApp: http://kb.vmware.com/kb/1010713 http://media.netapp.com/documents/tr-3749.pdf  HP EVA: http://h20195.www2.hp.com/v2/GetPDF.aspx/4AA1-2185ENW.pdf 64 Confidential
  • 65. Setting Round Robin Settings From CLI  Setting an entire SATP (Storage Array Type Plugin) to Round Robin: # esxcli nmp satp setdefaultpsp -s VMW_SATP_LSI -p VMW_PSP_RR  Setting the Round Robin PSP setting for a single LUN: # esxcli nmp device setpolicy -d naa.600174d001000000010f003048318438 --psp VMW_PSP_RR  Command line option to set iops value: # esxcli nmp roundrobin setconfig -d naa.600174d001000000010f003048318438 --iops 10 --type iops  Command line option to set bytes value: # esxcli nmp roundrobin setconfig -d naa.600174d001000000010f003048318438 --bytes 11 -- type bytes 65 Confidential
  • 66. Displaying Round Robin Settings From CLI  To displaying Round Robin settings for a LUN: # esxcli nmp device naa.60a9800050334b356b4a51312f417541 Device Display Name: NETAPP Fibre Channel Disk (naa.60a9800050334b356b4a51312f417541) Storage Array Type: VMW_SATP_ALUA Storage Array Type Device Config: {implicit_support=on;explicit_support=off;explicit_allow=on;alua_fol lowover=on;{TPG_id=2,TPG_state=AO}{TPG_id=3,TPG_state=ANO}} Path Selection Policy: VMW_PSP_RR Path Selection Policy Device Config: {policy=rr,iops=1000,bytes=10485760,useANO=0;lastPathIndex=3: NumIOsPending=0,numBytesPending=0} Working Paths: vmhba2:C0:T2:L1, vmhba1:C0:T2:L1 66 Confidential
  • 67. Known Issues for Round Robin in ESX 4.x  There are currently two known issues when using the Round Robin Path Selection Policy: • After setting the IOPS value for roundrobin, the setting is not retained after a reboot and instead shows a much higher value. This is resolved in a patch for ESX 4.0: http://kb.vmware.com/kb/1017721 • The SATP for EMC DMX (VMW_SATP_SYMM) does not distribute I/O evenly with roundrobin after a state change occurs. This state change includes adding a new path (FA) or when losing a paths and recovering. This is also resolved in a patch for ESX 4.0 and ESX 4.1: ESX 4.0: http://kb.vmware.com/kb/1023759 ESX 4.1: http://kb.vmware.com/kb/1027013 67 Confidential
  • 68. Agenda  vCenter Performance Charts & ESXTop  Disk Alignment  Hard Disks/Storage Array  SCSI Reservations  APD (All Paths Down) Issue  Multipathing Considerations  Communication 68 Confidential
  • 69. Communication 69 Confidential
  • 70. Communication  Communication between various business units is required for the success of a company.  Many companies silo their teams into networking, storage, server/hardware, Windows/Linux administrators, Database/Exchange administrators, virtualization administrators, and so on.  This silo design or syndrome has been responsible for lack of cooperation and communication between the siloed teams. There have been movements to try and break down these barriers however this is not an easy task.  Ultimately, the victims of this design may not be the people of the silo teams, but the business/company itself, which in turn affects customers directly. 70 Confidential
  • 71. Communication  In recent years, virtualization has become a huge movement in datacenters and the criticality of virtualized servers have shifted from simple application servers to mission critical applications. This mission critical requirement is far more demanding from an I/O (CPU, memory, and disk) and disk capacity perspective.  The need to have virtualization administrators and storage administrators communicate effectively has come to critical mass. More than ever, these two teams need to be on the same page at all times. The need to cross-pollinate knowledge between these teams has never been higher.  Lack of effective communication only between these two silo groups has led some companies to experience production outages and scalability issues. 71 Confidential
  • 72. Improving Communication  While change is not easy, there are some simple policies that can be adopted in order to break down the silo walls between the storage and virtualization administration teams: • Cross-train between both teams (basic to intermediate) • Meetings where topics that could affect the other silo should always have at least one representative from that silo that understands the discussion • Regular meetings (weekly) between the two silos to discuss current and future projects • Email distribution list that contains all parties of both silos • Always consider the other party before any changes are made to the virtualization infrastructure, array configuration, or policies on either team 72 Confidential
  • 73. Questions? 73 Confidential