2. Why a Management Cluster?
Dedicated resources for infrastructure management and
monitoring
Isolated from production & development workloads for
security and performance
Ability to troubleshoot production issues out-of-band
As vSphere environment grows, so do management and
monitoring needs
Getting vCloud Suite licensing – even more management
appliances
4/5/2016 2
Source: https://vvirtual.wordpress.com/2011/12/01/manage-your-virtual-environment-
wisely-separate-your-vmware-vsphere-management-servers-into-different-cluster/
3. Why VSAN?
Wanted dedicated storage for the management
cluster
◦ If production storage is having a bad day, tools to fix it
are still available
◦ Reduce I/O load on primary storage
Limited choices for dedicated arrays with decent
capacity & HA for small $$
VSAN integrated into vSphere – no additional
management tools
Simple
4/5/2016 3
5. Hardware
Targeted 2:1 vCPU:Logical Core ratio, 12 TB usable storage
Factored in current & future workloads plus organic growth
All components on VSAN hardware compatibility list
Modest IO needs, so Hybrid VSAN chosen over all-flash
Three Dell R730 2U servers, each with:
◦ 24 cores (2x Intel E5-2680 v3), 256 GB RAM
◦ 4x 10GbE (2x Intel X710)
◦ PERC H730 controller in HBA Mode
◦ 2x 400 GB SATA mix-use MLC SSD (cache)
◦ 8x 1 TB 7200 RPM NL-SAS (capacity)
24 TB raw capacity, 12 TB usable with VSAN FTT=1 (mirroring)
4/5/2016 5
Source: dell.com
6. Disk Layout
E S T
6 74 52 30 1 14 1512 1310 118 9
SAS
1TB7.2k
SAS
400GBSSD
SAS
400GBSSD
SAS
1TB7.2k
SAS
1TB7.2k
SAS
1TB7.2k
SAS
1TB7.2k
SAS
1TB7.2k
SAS
1TB7.2k
SAS
1TB7.2k
4/5/2016 6
Disk Group 1 Disk Group 2
Two Disk groups per host, each with 1x 400 GB SSD cache disk & 4x 1 TB HDD capacity disks
20% SSD to usable capacity ratio (best practice is 10% minimum)
Room to expand each disk group to 7 capacity disks (max for a VSAN disk group), or could add 3rd DG
7. Networking
4/5/2016 7
vmnic0
vmnic1 vmnic4
vmnic5
VMsManagement Replication ... iSCSI-2vMotion iSCSI-1 VSAN
Uplink 1 Uplink 2 Uplink 1 Uplink 2
Unused
Standby
Active
vmk0 vmk6 vmk2 vmk3 vmk4 vmk5
VDS-FrontEnd VDS-BackEnd
Port Group
Virtual Distributed Switch
Virtual Standard Switch
VMkernel NIC
8. Health Checks
VSAN Health Check Plugin integrated with vSphere 6.0 Update 1
◦ Checks health of all components
◦ Includes checking compatibility of hardware, drivers, firmware
◦ Proactive tests: VM creation, multicast performance, and storage performance
4/5/2016 8
9. Performance
Multiple test workloads available
Results using “70/30 read/write mix, realistic, optimal flash cache usage” test:
41,896 IOPS, 163.7 MB/sec throughput, 1.1ms avg. latency
4/5/2016 9
10. Issues/Challenges
Bug in 6.0 U1b:
◦ Warning shown, but all health checks are green
◦ KB 2143214 – cosmetic, resolved in 6.0 U2
◦ Can clear it by restarting management agents on hosts
Performance Monitoring:
◦ No performance charts in vSphere Web Client as of VSAN 6.1 (vSphere 6.0 U1) – new in 6.2
◦ Need to use VSAN Observer via RVC (Ruby vSphere Console) on vCenter to generate live perf stats
Multicast Performance Test:
◦ Only runs on 2 hosts in a 3-host cluster, shows lower than expected bandwidth, or fails
◦ Known issue, still present in 6.2
Bootstrap vCenter onto VSAN:
◦ vCenter needed to configure VSAN, so how do we get it there to start?
◦ Temporary iSCSI volume mounted to first host
4/5/2016 10
11. Tips
Check the HCL!
◦ Always important with vSphere, even more with VSAN
Read the VSAN 6.0 Design and Sizing Guide
◦ http://www.vmware.com/files/pdf/products/vsan/VSAN_Design_and_Sizing_Guide.pdf
Use the VSAN Sizing Calculator
◦ https://vsantco.vmware.com
Document your shutdown & startup procedures
◦ Potential chicken & egg problem with vCenter running on VSAN
◦ http://www.virtuallyghetto.com/2014/07/quick-tip-steps-to-shutdownstartup-vsan-cluster-wvcenter-
running-on-vsan-datastore.html
Set up alerts!
◦ Lots available, set notifications for anything health-related
4/5/2016 11