- Deduplication - A VMware datastore has a high level of data duplication. This is because it contains many instances of the same guest operating systems, application software, and so on. NetApp deduplication technology can eliminate this duplicated data, dramatically reducing the amount of storage required by your VMware environment. Space savings range from 50% to 90%, with 70% being typical.
 - Strongly consider using CAT6 cables rather than CAT5/5e. GbE will work on Cat5 cable and retransmissions will recover from any errors that occur, but they have a more significant impact on IP storage networks.
 - NFS NETWORKS AND STORAGE PERFORMANCE - There are 
three primary measures of storage performance: bandwidth (MBps), 
throughput in I/O operations per second (IOPS), and latency (ms). 
Throughput and bandwidth are related in the sense that the throughput 
measured in Mbps equals to IOPs multiplied by the I/O size. The I/O size
 is the size of the I/O operation from the host perspective.IOPS are 
usually determined by the back-end storage configuration. If the 
workload is cached, then it‘s determined by the cache response; most 
often, it‘s determined by the spindle configuration for the storage 
object. In the case of NFS datastores, the storage object is the file 
system. On a NetApp storage system, the IOPS achieved are primarily 
determined by the number of disk drives in an aggregate.
You should also know that each NFS datastore mounted by ESX 
(including vSphere) uses just one TCP session to an NFS datastore 
carrying both NFS control information and NFS data. Because of this, the
 upper limit throughput achievable from a single ESX host for a single 
datastore—regardless of link aggregation—is a single link. If you use 
1GbE, this means that a reasonable expectation is a unidirectional 
workload of ~80–100MB/sec (GbE is full duplex, so this can be 160MB/sec 
bidirectionally with a mixed workload). Higher total throughput on an 
ESX server can be achieved by leveraging multiple datastores. You can 
scale up the total throughput to multiple datastores using link 
aggregation and routing mechanisms.
The performance described above is sufficient for many use cases. Based on these numbers, NFS is appropriate for various VMware scenarios:
- A shared datastore supporting many VMs with an aggregate throughput requirement within the guidelines above (a large number of IOPS, but generally not large-block I/O)
 - A single busy VM as long as its I/O load can be served by a single GbE link
 
 - Partition Alignment - For NFS, there is no VMFS layer involved, so only the alignment of the guest VM file system within the VMDK to the NetApp storage array is required.The correct alignment for a new VM can be set using either diskpart to format a partition with the correct offset or by using fdisk from the ESX service console. In practice to avoid further creating of improperly aligned VMs, you must create templates that are properly aligned so new virtual machines built using those templates are properly aligned. NetApp has created a tool: mbralign, to check and correct the alignment of existing virtual machines.
 - THIN PROVISIONING - As mentioned above, the ability
 to take best advantage of thin provisioning is a major benefit of using
 NetApp NFS. VMware thin provisions the VMDK files on NFS datastores by 
default, but there are two types of thin-provisioned virtual disk files 
available:
- “Thick” type thin-provisioned virtual disk.This type of virtual disk
 file is created by default on NFS datastores during the virtual machine
 creation process. It has the following properties:
- Creates a flat .VMDK file; does not occupy actual disk blocks (thin provisioned) until there is a physical write from the guest OS
 - Guaranteed disk space reservation
 - Cannot oversubscribe the disk space on the NFS datastore
 
 - “Thin” type thin-provisioned virtual disk.You must create this type 
of virtual disk file using the vmkfstools command. It‘s properties are:
- Creates a flat .VMDK file; does not occupy actual disk blocks (thin provisioned) until there is a physical write from the guest OS
 - No guaranteed disk space reservation
 - Can oversubscribe the disk space on the NFS datastore
 
 
- # vdf -h /vmfs/volumes/<NFS Datastore Name>
 
Should an oversubscribed datastore encounter an out-of-space condition, all of the running VMs will become unavailable. The VMs simply “pause” waiting for space, but applications running inside of VMs may fail if the out-of-space condition isn’t addressed in a short period of time. For example, Oracle databases will remain active for 180 seconds; after that time has elapsed the database will fail. - “Thick” type thin-provisioned virtual disk.This type of virtual disk
 file is created by default on NFS datastores during the virtual machine
 creation process. It has the following properties:
 - High Availability and Disaster Recovery - NetApp 
recommends the following ESX failover timeout settings for NFS. We 
recommend increasing the default values to avoid VMs being disconnected 
during a failover event. NetApp VSC can configure these settings 
automatically. The settings that NetApp recommends (across all ESX 
hosts) are:
- NFS.HeartbeatFrequency (NFS.HeartbeatDelta in vSphere) = 12
 - NFS.HeartbeatTimeout = 5(default)
 - NFS.HeartbeatMaxFailures = 10. When the number of NFS datastores are
 increased, we also recommend increasing the heap values: 
Net.TcpipHeapSize =>’30′ to Net.TcpipHeapMax => ‘120′
- Back up your Windows registry.
 - Select Start>Run, type regedit.exe, and click OK.
 - In the left‐panel hierarchy view, double-click HKEY_LOCAL_MACHINE, then System, then CurrentControlSet, then Services, and then Disk.
 - Select the TimeOutValue and set the data value to 190 (decimal).
 
 
This means that the NFS datastore can be unavailable for a maximum of 125 seconds before being marked unavailable, which covers the large majority of NetApp failover events.
During this time period, a guest sees a non-responsive SCSI disk on the vSCSI adapter. The disk timeout is how long the guest OS will wait when the disk is non-responsive. Use the following procedure to set operating system timeouts for Windows servers to match the 190-second maximum set for the datastore:
- Back up your Windows registry.
 - Select Start>Run, type regedit.exe, and click OK.
 - In the left‐panel hierarchy view, double-click HKEY_LOCAL_MACHINE, then System, then CurrentControlSet, then Services, and then Disk.
 - Select the TimeOutValue and set the data value to 190 (decimal).
 
 
Monday, August 13, 2012
VMware on NetApp NFS
This information was taken from a NetApp white paper by Bikash 
Roy Choudhury – TR-3839.  I am just pulling out snippets that I thought 
were very interesting.  Not all of it is specific to NetApp, but most of
 my consulting work is on NetApp.
Labels:
Disaster Recovery,
High Availability,
NetApp,
NFS,
VMware,
Windows
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment
Thanks for your comment!