Beruflich Dokumente
Kultur Dokumente
NFS & iSCSI: Performance Characterization and Best Practices in ESX 3.5
Priti Mishra
MTS, VMware
Bing Tsai
Sr. R&D Manager, VMware
Housekeeping
Please turn off your mobile phones, blackberries and laptops Your feedback is valued: please fill in the session evaluation form (specific to that session) & hand it to the room monitor / the materials pickup area at registration Each delegate to return their completed event evaluation form to the materials pickup area will be eligible for a free evaluation copy of VMwares ESX 3i Please leave the room between sessions, even if your next session is in the same room as you will need to be rescanned
Topics
General Performance Data and Comparison
Improvements in ESX 3.5 over ESX 3.0.x
Access mode
50% read/ write
Access pattern
100% sequential
Cached runs
100MB data disks to minimize array/server disk activities All I/Os served from server/array cache Gives upper bound on performance
Networking Configuration
Dedicated VLANs for data traffic isolated from general networking
Latency
Lower is better Negative is better lower response time
CPU cost
Lower is better Negative is better reduced CPU cost How does this metric matter?
CPU Costs
Why is CPU cost data useful?
Determines how much I/O traffic the system CPUs can handle
How many I/O-intensive VMs can be consolidated in a host
Performance Data
First set: Relative to baselines in ESX 3.0.x Second set: Comparison of storage options using Fibre Channel data as the baseline Last: VMFS vs. RDM physical
Se uentia
ifference
3 I Si e ( yte)
Latency is lower
Especially for smaller data sizes Read operations benefit most
Throughput levels
Dependent on workload
Mixed read-write patterns show most gain Read I/Os show gains for small data sizes
Se uentia
ifference
Si e ( yte)
Latency is better
Smaller data sizes show the most gain Mixed read-write and read I/Os benefit more
Throughput levels
Dependent on workload
Mixed read-write patterns show most gain for all block sizes Pure read and write I/Os show gains for small block sizes
Protocol Comparison
Which storage option to choose?
IP Storage vs. Fibre Channel
u ntia -
rit
1.4
H/ i C i C
1.2
/
1.0 0.8 0.6 0.4 0.2 0.0 1k 4k 8k 16k 32k 64k 72k
128k
256k
512k
Comparison with FC -
esponse ime - e
entia
rite
H/ / F
i C i C
1.5
1.0
0.5
tial
tial
25
Topics
General Performance Data and Comparison
Improvements in ESX 3.5 over ESX 3.0.x
verview
Additional Considerations (1 of 3)
NFS parameters # of NFS mount points
Multiple VMs using multiple mount points may give higher aggregate throughput with slightly higher CPU cost
Export option on NFS server affects performance iSCSI protocol parameters Header digest processing: slight impact on performance Data digest processing: turning off may result in
Improved CPU utilization Slightly lower latencies Minor throughput improvement Actual outcome highly dependent on workload
Additional Considerations (2 of 3)
NUMA specific
If only one VM is doing heavy I/O, may be beneficial to pin the VM and its memory to node 0 If CPU usage is not a concern; no pinning necessary On each VM reboot, ESX Server will place it on the next adjacent NUMA node Minor performance implications for certain workloads To avoid this movement, VM should be affinitized using VI client
SMP VMs
For I/O workloads within an SMP VM that migrate frequently between VCPUs Pin the guest thread/process to a specific VCPU Some versions of Linux has KHz timer rate and may incur high overhead
Additional Considerations (3 of 3)
CPU headroom Software initiated iSCSI and NFS protocols can consume significant amount of CPU in certain I/O patterns
Small I/O workloads require large amount of CPU; ensure that CPU saturation does not restrict I/O rate
Networking Avoid link over-subscription Ensure all networking parameters or even the basic gigabit connection is consistent across the full network path Intelligent use of VLAN or zoning to minimize traffic interference
From esxtop data CPU utilization also lower for l/O sizes larger than 128KB CPU cost per I/O is in expected range for all I/O sizes
But very little difference in performance between random and sequential reads From NFS server spec 3GB read/write cache
Most data should be in cache after warming up
Random write latency in the 20ms range Sequential write < 1ms From NFS server stats cache hit% much lower for random writes, even after warm-up
From native tests Random and sequential write performance is almost same From network packet traces Server responds to random writes in 10 to 20ms, sequential writes in <1ms Offset in NFS WRITE requests is not aligned to power-of-2 boundary
Packet traces from native runs show correct alignment
Solution
Use disk alignment tool in the guest OS to align disk partition Alternatively, use unformatted partition inside guest OS
Questions?