Vmware Vsan 8 0
Vmware Vsan 8 0
0
VMware vSAN 8.0
Table of Contents
Release Notes.....................................................................................................................................12
VMware vSAN 8.0 Update 3 Release Notes................................................................................................................ 12
VMware vSAN 8.0 Update 2 Release Notes................................................................................................................ 16
VMware vSAN 8.0 Update 1 Release Notes................................................................................................................ 26
VMware vSAN 8.0 Release Notes.................................................................................................................................38
vSAN Planning and Deployment...................................................................................................... 50
Updated Information...................................................................................................................................................... 50
What Is vSAN................................................................................................................................................................ 50
vSAN Concepts......................................................................................................................................................... 51
Characteristics of vSAN.....................................................................................................................................51
vSAN Terms and Definitions..............................................................................................................................53
How vSAN Differs from Traditional Storage...................................................................................................... 56
Building a vSAN Cluster............................................................................................................................................... 56
vSAN Deployment Options....................................................................................................................................... 58
Integrate vSAN with Other VMware Software............................................................................................................. 59
Limitations of vSAN....................................................................................................................................................... 60
Requirements for Enabling vSAN................................................................................................................................ 60
Hardware Requirements for vSAN.......................................................................................................................... 60
Cluster Requirements for vSAN................................................................................................................................62
Software Requirements for vSAN.............................................................................................................................62
Networking Requirements for vSAN......................................................................................................................... 62
License Requirements...............................................................................................................................................62
Per TiB License for vSAN................................................................................................................................. 63
VMware Cloud Foundation License for vSAN...................................................................................................63
VMware vSphere Foundation Capacity License for vSAN................................................................................63
Per CPU License for vSAN............................................................................................................................... 64
Per Core License for vSAN............................................................................................................................... 64
Designing and Sizing a vSAN Cluster......................................................................................................................... 65
Designing and Sizing vSAN Storage........................................................................................................................65
Planning Capacity in vSAN............................................................................................................................... 65
Design Considerations for Flash Caching Devices in vSAN.............................................................................67
Design Considerations for Flash Capacity Devices in vSAN............................................................................ 68
Design Considerations for Magnetic Disks in vSAN......................................................................................... 69
Design Considerations for Storage Controllers in vSAN...................................................................................70
Designing and Sizing vSAN Hosts........................................................................................................................... 70
Design Considerations for a vSAN Cluster.............................................................................................................. 71
3
VMware vSAN 8.0
4
VMware vSAN 8.0
5
VMware vSAN 8.0
6
VMware vSAN 8.0
7
VMware vSAN 8.0
8
VMware vSAN 8.0
9
VMware vSAN 8.0
10
VMware vSAN 8.0
11
VMware vSAN 8.0
Release Notes
Release notes include product enhancements and notices, bug fixes, and resolved issues.
Introduction
What's New
vSAN 8.0 Update 3 introduces the following new features and enhancements:
Licensing
Capacity-based licensing. The subscription-based licensing in VMware Cloud Foundation 5.2 entitles customers to 1
Tebibyte (TiB) of vSAN capacity per VCF core license, with additional capacity available through a capacity-based add-on
license.
Flexible Topologies
Stretched cluster support on vSAN ESA. In VMware Cloud Foundation 5.2 vSAN Express Storage Architecture fully
supports stretched cluster topologies. This enables VCF to configure workload and management domains that provide
site-level resilience for your workloads and data.
vSAN Max as principal storage. VCF 5.2 supports vSAN Max as primary, centralized shared storage. This enables VCF
to configure workload domains with disaggregated vSAN Max clusters, in addition to aggregated vSAN HCI clusters, to
increase your flexibility.
vSAN File Services up to 250 file shares. vSAN 8.0 Update 3 improves the scalability of native File Services in VMware
Cloud Foundation by increasing the number of shares per cluster to 250 on vSAN ESA.
Data Protection
12
VMware vSAN 8.0
vSAN local data protection leveraging ESA scalable snapshots. vSAN data protection enables you to capture local
snapshots using an intuitive new UI, and store them on your vSAN datastore. Use protection groups to easily define VM
membership, snapshot schedules, retention, and immutability criteria for VMs. You can use these snapshots to revert,
restore, recover, or clone VMs for enhanced levels of protection.
Integration with VMware Live Cyber Recovery (VLCR), for faster ransomware recovery. vSAN data protection
integration with VMware Live Cyber Recovery (VLCR), captures point-in-time snapshots for cloud-based ransomware
protection, and VLCR provides the tools to protect data off-site and recover VMs in an isolated recovery environment
(IRE) for analysis before restoring them on-premises. VLCR uses a local snapshot, and only updates the changes (deltas)
from IRE to production, drastically reducing restore times.
Improved Resiliency
Congestion remediation. vSAN 8.0 Update 3 enhances vSAN OSA's ability to detect and remediate various types of
congestion early, preventing cluster-wide I/O latencies.
Adaptive delete congestion. vSAN now provides adaptive delete congestion for compression-only disk groups in vSAN
OSA, improving IOPS performance and delivering more predictable application responses.
Enhanced Management
Proactive hardware management. vSAN 8.0 Update 3 introduces a new method for collecting critical storage device
telemetry from preferred server vendors, enabling predictive management of hardware issues. Proactive Hardware
Management leverages OEM vendors' predictive failure monitoring tools integrated into vCenter via API (after OEM
Hardware Support Manager is setup and configured), to help you make more informed decisions about hardware
maintenance.
Data-at-rest encryption disable for vSAN ESA. You can disable data-at-rest encryption on vSAN ESA clusters at any
point after enabling it. vSAN ESA now supports the following operations for data-at-rest encryption: enable encryption,
disable encryption, shallow rekey, and deep rekey.
Customizable alarm thresholds for NVMe storage devices in vSAN ESA. vSAN 8.0 Update 3 enables you to
customize alert thresholds for device endurance, tailoring them to specific clusters, hosts, disk vendors, or even individual
devices.
vSAN I/O Trip Analyzer cluster level view. vSAN 8.0 Update 3 enables you to run performance analysis on multiple
VMs simultaneously. You can select up to 8 VMs at once, and quickly perform analysis on each selected VM.
Enhanced awareness of vSAN Max using Aria Operations. VMware Cloud Foundation Operations with VCF 5.2
introduces enhanced visibility for vSAN Max clusters throughout its user interface. This enables Aria Operations to track
resource utilization and health status.
Federated vSAN health monitoring in Aria Operations. The latest Aria Operations introduces federated vSAN cluster
health monitoring for clusters spanning across multiple vCenters.
Deprecation of locales:
13
VMware vSAN 8.0
Beginning with the next major release, we will be reducing the number of supported localization languages. The three
supported languages will be:
• Japanese
• Spanish
• French
The following languages will no longer be supported:
• Italian, German, Brazilian Portuguese, Traditional Chinese, Korean, Simplified Chinese
Impact:
• Users who have been using the deprecated languages will no longer receive updates or support in these languages.
• All user interfaces, help documentation, and customer support will be available in the three supported languages
mentioned above.
14
VMware vSAN 8.0
Limitations
For information about maximum configuration limits for the vSAN 8.0 Update 3 release, see the Configuration Maximums
documentation.
Known Issues
Workaround: To release the storage back to vSAN immediately, delete the file shares.
15
VMware vSAN 8.0
Introduction
What's New
vSAN 8.0 Update 2 introduces the following new features and enhancements:
16
VMware vSAN 8.0
Disaggregated Storage
Enhanced topologies for disaggregation with vSAN Express Storage Architecture bring feature parity for vSAN OSA and
vSAN ESA.
vSAN ESA support for stretched clusters in disaggregated topology. vSAN ESA supports disaggregation when
using vSAN stretched clusters. In addition to supporting several stretched cluster configurations, vSAN also optimizes the
network paths for certain topologies to improve the performance capabilities of stretched cluster configurations.
Support of disaggregation across clusters using multiple vCenter Servers. vSAN 8.0 Update 2 supports
disaggregation across environments using multiple vCenter Servers when using vSAN ESA. This enables vSphere or
vSAN clusters managed by one vCenter Server to use the storage resources of a vSAN cluster managed by a different
vCenter Server..
vSAN ESA Adaptive Write path for disaggregated storage. Disaggregated deployments get the performance benefits
of a new adaptive write path previously introduced in vSAN 8.0 Update 1 for standard ESA based deployments. VMs
running on a vSphere or vSAN cluster that consume storage from another vSAN ESA cluster can take advantage of this
capability. Adaptive write path technology in a disaggregated environment helps your VMs achieve higher throughput and
lower latency, and do so automatically in real time, without any interaction by the administrator.
Core Platform Enhancements
Integrated File Services for Cloud Native and traditional workloads. vSAN 8.0 Update 2 supports vSAN File Service
on vSAN Express Storage Architecture. File service clients can benefit from performance and efficiency enhancements
provided by vSAN ESA.
Adaptive Write Path optimizations in vSAN ESA. vSAN ESA introduces an adaptive write path that helps the cluster
ingest and process data more quickly. This optimization improves performance for workloads driving high I/O to single
object (VMDK), and also improves aggregate cluster performance.
Increased number of VM's per host in vSAN ESA clusters (up to 500/host). vSAN 8.0 Update 2 supports up to 500
VMs per host VM on vSAN ESA clusters, provided the underlying hardware infrastructure can support it. Now you can
leverage NVMe-based high performance hardware platforms optimized for the latest generation of CPUs with high core
densities, and consolidate more VMs per host.
New ReadyNode profile and support for read-intensive devices for vSAN ESA. vSAN ESA announces the availability
of new ReadyNode profiles designed for small data centers and edge environments with lower overall hardware
requirements on a per-node basis. This release also introduces support for read-intensive storage devices.
vSAN ESA support for encryption deep rekey. vSAN clusters using data-at-rest encryption have the ability to perform a
deep rekey operation. A deep rekey decrypts the data that has been encrypted and stored on a vSAN cluster using the old
encryption key, and re-encrypts the data using newly issued encryption keys prior to storing it on the vSAN cluster.
Enriched Operations
vSAN ESA prescriptive disk claim. vSAN ESA includes a prescriptive disk claim process that further simplifies
management of storage devices in each host in a vSAN cluster. This feature provides consistency to the disk claiming
process during initial deployment and cluster expansion.
Capacity reporting enhancements. Overhead breakdown in vSAN ESA space reporting displays both the ESA object
overhead and the original file system overhead.
Auto-Policy management improvements in vSAN ESA. Enhanced auto-policy management feature determines if the
default storage policy needs to be adjusted when a user adds or removes a host from a cluster. If vSAN identifies a need
to change the default storage policy, it triggers a health check warning. You can make the change with a simple click at
which time vSAN reconfigures the cluster with the new policy.
Skyline Health remediation enhancements. vSAN Skyline Health helps you reduce resolution times by providing
deployment-specific guidance along with more prescriptive guidance on how to resolve issues.
17
VMware vSAN 8.0
Key expiration for clusters with data-at-rest encryption. vSAN 8.0 Update 2 supports the use of KMS servers with a
key expiration attribute used for assigning an expiration date to a Key Encryption Key (KEK).
I/O top contributors enhancements. vSAN Performance Service has improved the process to find performance hot
spots over a customizable time period to help you diagnose performance issues while using multiple types of sources for
analysis (VMs, host disks, and so on).
I/O Trip Analyzer supported on two node clusters and stretched clusters. vSAN 8.0 Update 2 has enhanced the I/
O Trip Analyzer to report on workloads in a vSAN stretched cluster. Now you can determine where the primary source of
latency is occurring in a vSAN stretched cluster, as well as latencies in other parts of the stack that can contribute to the
overall latency experienced by the VM.
Easier configuration for two node clusters and stretched clusters. Several new features to help management of two
node and stretched cluster deployments.
• Witness host traffic configured in the vSphere Client.
• Support for medium sized witness host appliance in vSAN ESA.
• Support in vLCM to manage lifecycle of shared witness host appliance types.
18
VMware vSAN 8.0
Disk format version 1.0 does not have performance and snapshot enhancements, and it lacks support for advanced
features including checksum, deduplication and compression, and encryption. For more information about vSAN disk
format version, see KB 2148493.
Upgrading the On-disk Format for Hosts with Limited Capacity
During an upgrade of the vSAN on-disk format from version 1.0 or 2.0, a disk group evacuation is performed. The disk
group is removed and upgraded to on-disk format version 17.0, and the disk group is added back to the cluster. For two-
node or three-node clusters, or clusters without enough capacity to evacuate each disk group, select Allow Reduced
Redundancy from the vSphere Client. You also can use the following RVC command to upgrade the on-disk format:
vsan.ondisk_upgrade --allow-reduced-redundancy
When you allow reduced redundancy, your VMs are unprotected for the duration of the upgrade, because this method
does not evacuate data to the other hosts in the cluster. It removes each disk group, upgrades the on-disk format, and
adds the disk group back to the cluster. All objects remain available, but with reduced redundancy.
If you enable deduplication and compression during the upgrade, you can select Allow Reduced Redundancy from the
vSphere Client.
Limitations
For information about maximum configuration limits for the vSAN 8.0 Update 2 release, see the Configuration Maximums
documentation.
Known Issues
19
VMware vSAN 8.0
4. Merge the two lines into one as shown below, and save the file.
5. Restart the vSAN health service: /usr/sbin/vmon-cli -r vsan-health
Before:
vsan.health.optimaldsdefaultpolicy.htf.0.default.0.des=健全性テーブルに表示される
vSAN の最適なデータストアのデフォルト ポリシーの「許容される障害の数」または「サイトの耐障害性」の構成 (ある
いは両方) が推奨ポリシーと一致しません。
After:
vsan.health.optimaldsdefaultpolicy.htf.0.default.0.des=健全性テーブルに表示される vSAN の最適なデータストアのデ
フォルト ポリシーの「許容される障害の数」または「サイトの耐障害性」の構成 (あるいは両方) が推奨ポリシーと一致
しません。
hostAffinity policy option lost during upgrade
When you upgrade from vSAN 6.7 to vSAN 8.0, the vCenter Server hostaffinity option value is changed to false.
Workaround: Set the hostaffinity option back to true to continue using vSAN HostLocal policy for a normal VM.
Cannot enable File Service if vCenter Server internet connectivity is disabled
If you disable vCenter Server internet connectivity, the Enable File Service dialog does not display File service agent
section and you cannot select OVF.
Workaround: To enable vCenter Server internet connectivity:
1. Navigate to Cluster > Configure > vSAN > Internet Connectivity.
2. Click Edit to open Edit Internet Connectivity dialog.
3. Select Enable Internet access for all vSAN clusters checkbox and click Apply.
Cannot deactivate encryption on vSAN ESA
After you enable data-at-rest encryption on a vSAN ESA cluster, you cannot deactivate it.
Workaround: None.
vSAN File Service does not support NFSv4 delegations
vSAN File Service does not support NFSv4 delegations in this release.
Workaround: None.
In stretched cluster, file server with no affinity cannot rebalance
In the stretched cluster vSAN File Service environment, a file server with no affinity location configured cannot be
rebalanced between Preferred ESXi hosts and Non-preferred ESXi hosts.
Workaround: Set the affinity location of the file server to Preferred or Non-Preferred by editing the file service domain
configuration.
Remediation of ESXi hosts in a vSphere Lifecycle Manager cluster with vSAN fails if vCenter services are
deployed on custom ports
If vCenter Server services are deployed on custom ports in a cluster with vSAN, vSphere DRS, and vSphere HA,
remediation of vSphere Lifecycle Manager clusters might fail. This problem is caused by a vSAN resource health check
error. ESXi hosts cannot enter maintenance mode, which leads to failing remediation tasks.
Workaround: None.
When vSAN file service is enabled, DFC-related operations such as upgrade, enabling encryption or data-
efficiency might fail
20
VMware vSAN 8.0
When file service is enabled, an agent VM runs on each host. The underlying vSAN object might be placed across
multiple diskgroups. When the first diskgroup gets converted, the vSAN object becomes inaccessible and the agent VM
is in an invalid state. If you try to delete the VM and redeploy a new VM, the operation fails due to the VM’s invalid state.
The VM gets unregistered but the inaccessible object still exists there. When the next diskgroup gets converted, there is
a precheck for inaccessible objects in the whole cluster. This check fails the DFC since it finds inaccessible objects of the
old agent VM.
Workaround: Manually remove the inaccessible objects.
When such failure happens, you can see the DFC task failure.
1. Identify the inaccessible objects from the failure task fault information.
2. To ensure that the objects belong to the agent VM, inspect the hostd log file and confirm that the objects belong to the
VM’s object layout.
3. Log in to the host and use the /usr/lib/vmware/osfs/bin/objtool command to remove the objects manually.
Note: To prevent this problem, disable file service before performing any DFC-related operation.
Workaround: To release the storage back to vSAN immediately, delete the file shares.
21
VMware vSAN 8.0
22
VMware vSAN 8.0
When a vSAN cluster is full with one or more disk groups reaching 100%, there can be a VM pending question that
requires user action. If the question is not answered and if the cluster full condition is left unattended, the IP addresses
VMs might change to IPv6 or become unavailable. This prevents you from using SSH to access the VMs. It also prevents
you from using the VM console, because the console goes blank after you type root .
Workaround: None.
Unable to remove a dedupe enabled disk group after a capacity disk enters PDL state
When a capacity disk in a dedupe-enabled disk group is removed, or its unique ID changes, or when the device
experiences an unrecoverable hardware error, it enters Permanent Device Loss (PDL) state. If you try to remove the disk
group, you might see an error message informing you that the action cannot be completed.
Workaround: Whenever a capacity disk is removed, or its unique ID changes, or when the device experiences an
unrecoverable hardware error, wait for a few minutes before trying to remove the disk group.
In deduplication clusters, reactive rebalancing might not happen when the disks show more than 80% full
In deduplication clusters, when the disks display more than 80% full on the dashboard, the reactive rebalancing might not
start as expected. This is because in deduplication clusters, pending writes and deletes are also considered for calculating
the free capacity.
Workaround: None.
TRIM/UNMAP commands from Guest OS fail
If the Guest OS attempts to perform space reclamation during online snapshot consolidation, the TRIM/UNMAP
commands fail. This failure keeps space from being reclaimed.
Workaround: Try to reclaim the space after the online snapshot operation is complete. If subsequent TRIM/UNMAP
operations fail, remount the disk.
Space reclamation from SCSI TRIM/UNMAP is lost when online snapshot consolidation is performed
Space reclamation achieved from SCSI TRIM/UNMAP commands is lost when you perform online snapshot consolidation.
Offline snapshot consolidation does not affect SCSI unmap operation.
Workaround: Reclaim the space after online snapshot consolidation is complete.
23
VMware vSAN 8.0
Workaround: Wait until no resynchronization is pending, then retry the disk format upgrade.
Powered off VMs appear as inaccessible during witness host replacement
When you change a witness host in a stretched cluster, VMs that are powered off appear as inaccessible in the vSphere
Web Client for a brief time. After the process is complete, powered off VMs appear as accessible. All running VMs appear
as accessible throughout the process.
Workaround: None.
Cannot place hosts in maintenance mode if they have faulty boot media
vSAN cannot place hosts with faulty boot media into maintenance mode. The task to enter maintenance mode might fail
with an internal vSAN error, due to the inability to save configuration changes. You might see log events similar to the
following: Lost Connectivity to the device xxx backing the boot filesystem
Workaround: Remove disk groups manually from each host, using the Full data evacuation option. Then place the host
in maintenance mode.
After stretched cluster failover, VMs on the preferred site register alert: Failed to failover
If the secondary site in a stretched cluster fails, VMs failover to the preferred site. VMs already on the preferred site might
register the following alert: Failed to failover.
Workaround: Ignore this alert. It does not impact the behavior of the failover.
During network partition, components in the active site appear to be absent
During a network partition in a vSAN two-host or stretched cluster, the vSphere Web Client might display a view of the
cluster from the perspective of the non-active site. You might see active components in the primary site displayed as
absent.
Workaround: Use RVC commands to query the state of objects in the cluster. For example: vsan.vm_object_info
24
VMware vSAN 8.0
When you move a host from an encrypted vSAN cluster to another encrypted vSAN cluster, then move the host back to
the original encrypted cluster, the task might fail. You might see the following message: A general system error
occurred: Invalid fault . This error occurs because vSAN cannot re-encrypt data on the host using the original
encryption key. After a short time, vCenter Server restores the original key on the host, and all unmounted disks in the
vSAN cluster are mounted.
Workaround: Reboot the host and wait for all disks to get mounted.
Cannot perform deep rekey if a disk group is unmounted
Before vSAN performs a deep rekey, it performs a shallow rekey. The shallow rekey fails if an unmounted disk group is
present. The deep rekey process cannot begin.
Workaround: Remount or remove the unmounted disk group.
Log entries state that firewall configuration has changed
A new firewall entry appears in the security profile when vSAN encryption is enabled: vsanEncryption. This rule controls
how hosts communicate directly to the KMS. When it is triggered, log entries are added to /var/log/vobd.log . You
might see the following messages:
Firewall configuration has changed. Operation 'addIP4' for rule set vsanEncryption
succeeded.
Firewall configuration has changed. Operation 'removeIP4' for rule set vsanEncryption
succeeded.
These messages can be ignored.
Workaround: None.
HA failover does not occur after setting Traffic Type option on a vmknic to support witness traffic
If you set the traffic type option on a vmknic to support witness traffic, vSphere HA does not automatically discover the
new setting. You must manually disable and then re-enable HA so it can discover the vmknic. If you configure the vmknic
and the vSAN cluster first, and then enable HA on the cluster, it does discover the vmknic.
Workaround: Manually disable vSphere HA on the cluster, and then re-enable it.
After resolving network partition, some VM operations on linked clone VMs might fail
Some VM operations on linked clone VMs that are not producing I/O inside the guest operating system might fail. The
operations that might fail include taking snapshots and suspending the VMs. This problem can occur after a network
partition is resolved, if the parent base VM's namespace is not yet accessible. When the parent VM's namespace
becomes accessible, HA is not notified to power on the VM.
Workaround: Power cycle VMs that are not actively running I/O operations.
Cannot place a witness host in Maintenance Mode
When you attempt to place a witness host in Maintenance Mode, the host remains in the current state and you see the
following notification: A specified parameter was not correct.
Workaround: When placing a witness host in Maintenance Mode, choose the No data migration option.
Moving the witness host into and then out of a stretched cluster leaves the cluster in a misconfigured state
If you place the witness host in a vSAN-enabled vCenter cluster, an alarm notifies you that the witness host cannot reside
in the cluster. But if you move the witness host out of the cluster, the cluster remains in a misconfigured state.
Workaround: Move the witness host out of the vSAN stretched cluster, and reconfigure the stretched cluster. For more
information, see this article: https://kb.vmware.com/s/article/2130587.
25
VMware vSAN 8.0
Unmounted vSAN disks and disk groups displayed as mounted in the vSphere Web Client Operational Status
field
After the vSAN disks or disk groups are unmounted by either running the esxcli vsan storage disk group
unmount command or by the vSAN Device Monitor service when disks show persistently high latencies, the vSphere
Web Client incorrectly displays the Operational Status field as mounted.
Workaround: Use the Health field to verify disk status, instead of the Operational Status field.
What's New
vSAN 8.0 Update 1 introduces the following new features and enhancements:
Disaggregated Storage
Disaggregation with vSAN Express Storage Architecture. vSAN 8.0 Update 1 provides disaggregation support for
vSAN Express Storage Architecture (ESA), as it is supported with vSAN Original Storage Architecture (OSA). You can
mount remote vSAN datastores that reside in other vSAN ESA server clusters. You also can use an ESA cluster as the
external storage resource for a compute-only cluster. All capabilities and limits that apply to disaggregation support for
vSAN OSA also apply to vSAN ESA. vSAN ESA client clusters can connect only to a vSAN ESA based server cluster.
Disaggregation for vSAN stretched clusters (vSAN OSA). This release supports vSAN stretched clusters in
disaggregated topology. In addition to supporting several stretched cluster configurations, vSAN can optimize network
paths for certain topologies to improve stretched cluster performance.
Disaggregation across clusters using multiple vCenter Servers (vSAN OSA). vSAN 8.0 Update 1 introduces support
for vSAN OSA disaggregation across environments using multiple vCenter Servers. This enables clusters managed by
one vCenter Server to use storage resources that reside on a vSAN cluster managed by a different vCenter Server.
26
VMware vSAN 8.0
host in addition to the hosts holding the data. This helps ensure the durability of the changed data if additional hosts
fail while the original host is still in maintenance mode.
• Increased administrative storage capacity on vSAN datastores using customizable namespace objects. You
can customize the size of namespace objects that enable administrators to store ISO files, VMware content library, or
other infrastructure support files on a vSAN datastore.
• Witness appliance certification. In vSAN 8.0 Update 1, the software acceptance level for vSAN witness appliance
has changed to Partner Supported. All vSphere Installation Bundles (VIBs) must be certified.
Simplified Management
• Auto-policy management for the default storage policy (vSAN ESA). vSAN ESA introduces auto-policy
management, an optional feature that creates and assigns a default storage policy designed for the cluster. Based on
the size and type of cluster, auto-policy management selects the ideal level of failure to tolerate and data placement
scheme. Skyline health uses this data to monitor and alert you if the default storage policy is ideal or sub-optimal, and
guides you to adjust the default policy based on the cluster characteristics. Skyline health actively monitors the cluster
as its size changes, and provides new recommendations as needed.
• Skyline health intelligent cluster health scoring, diagnostics and remediation. Improve efficiency by using the
cluster health status and troubleshooting dashboard that prioritizes identified issues, enabling you to focus and take
action on the most important issues.
• High resolution performance monitoring in vSAN performance service. vSAN performance service provides real-
time monitoring of performance metrics that collects and renders metrics every 30 seconds, making monitoring and
troubleshooting more meaningful. VMware snapshot APIs are unchanged. VMware VADP supports all vSAN ESA
native snapshot operations on the vSphere platform.
• VM I/O trip analyzer task scheduling. VM I/O trip analyzer can schedule based on time-of-day, for a particular
duration and frequency to capture details for repeat-offender VMs. The diagnostics data collected are available for
analysis in the VM I/O trip analyzer interface in vCenter.
• PowerCLI enhancements. PowerCLI supports the following new capabilities:
vSAN ESA disaggregation
vSAN OSA disaggregation for stretched clusters
vSAN OSA disaggregation across multiple vCenter Servers
vSAN cluster shutdown
Object format updates and custom namespace objects
27
VMware vSAN 8.0
Note: Before performing the upgrade, please review the most recent version of the VMware Compatibility Guide to
validate that the latest vSAN version is available for your platform.
Note: vSAN Express Storage Architecture is available only for new deployments. You cannot upgrade a cluster to vSAN
ESA.
vSAN 8.0 Update 1 is a new release that requires a full upgrade to vSphere 8.0 Update 1. Perform the following tasks to
complete the upgrade:
1. Upgrade to vCenter Server 8.0 Update 1. For more information, see the VMware vSphere 8.0 Update 1 Release
Notes.
2. Upgrade hosts to ESXi 8.0 Update 1. For more information, see the VMware vSphere 8.0 Update 1 Release Notes.
3. Upgrade the vSAN on-disk format to version 18.0. If upgrading from on-disk format version 3.0 or later, no data
evacuation is required (metadata update only).
4. Upgrade FSVM to enable new File Service features and get all the latest updates.
Note: vSAN retired disk format version 1.0 in vSAN 7.0 Update 1. Disks running disk format version 1.0 are no
longer recognized by vSAN. vSAN will block upgrade through vSphere Update Manager, ISO install, or esxcli to vSAN 7.0
Update 1. To avoid these issues, upgrade disks running disk format version 1.0 to a higher version. If you have disks on
version 1.0, a health check alerts you to upgrade the disk format version.
Disk format version 1.0 does not have performance and snapshot enhancements, and it lacks support for advanced
features including checksum, deduplication and compression, and encryption. For more information about vSAN disk
format version, see KB 2148493.
Upgrading the On-disk Format for Hosts with Limited Capacity
During an upgrade of the vSAN on-disk format from version 1.0 or 2.0, a disk group evacuation is performed. The disk
group is removed and upgraded to on-disk format version 17.0, and the disk group is added back to the cluster. For two-
node or three-node clusters, or clusters without enough capacity to evacuate each disk group, select Allow Reduced
Redundancy from the vSphere Client. You also can use the following RVC command to upgrade the on-disk format:
vsan.ondisk_upgrade --allow-reduced-redundancy
When you allow reduced redundancy, your VMs are unprotected for the duration of the upgrade, because this method
does not evacuate data to the other hosts in the cluster. It removes each disk group, upgrades the on-disk format, and
adds the disk group back to the cluster. All objects remain available, but with reduced redundancy.
If you enable deduplication and compression during the upgrade, you can select Allow Reduced Redundancy from the
vSphere Client.
Limitations
For information about maximum configuration limits for the vSAN 8.0 Update 1 release, see the Configuration Maximums
documentation.
Known Issues
28
VMware vSAN 8.0
This issue affects compute-only client clusters that mount from a vSAN ESA server cluster. When you mount a remote
datastore, the datastore capacity value shown on the host and the vSphere client does not match the actual value. Aside
from the reporting issue, there is no known impact on VM operations.
Workaround: None.
Cannot enable File Service if vCenter Server internet connectivity is disabled
If you disable vCenter Server internet connectivity, the Enable File Service dialog does not display File service agent
section and you cannot select OVF.
Workaround: To enable vCenter Server internet connectivity:
1. Navigate to Cluster > Configure > vSAN > Internet Connectivity.
2. Click Edit to open Edit Internet Connectivity dialog.
3. Select Enable Internet access for all vSAN clusters checkbox and click Apply.
KMS connection health checks not available when KMS is offline
This issue affects vSAN health checks for clusters with data-at-rest encryption. When the KMS is offline, the following
health check might not be available: VMware vCenter and all hosts are connected to Key Management Servers. If this
issue occurs, you cannot see warnings or errors that indicate the offline status of the KMS.
Workaround: None
Mount remote datastore from a stretched server cluster fails with message: Site affinity provided in
server cluster configuration are not present
Mounting a remote datastore from a stretched server cluster might fail under the following conditions:
• The client vSAN cluster already has another datastore from a different stretched server cluster.
• The stretched server clusters have different fault domain names.
• The client has asymmetric network topology to both the server clusters.
The following message is displayed: Site affinity provided in server cluster configuration are not
present in server cluster fault domains
Workaround: Rename the fault domains on both server clusters to match, and retry the operation.
Sequential workload performance improvements not enabled
Some performance improvements for sequential workloads cannot take effect until the vSAN object moves from the
host or the host is rebooted. You must manually abdicate the DOM owner of all vSAN objects to enable performance
improvements.
Workaround: After you upgrade from vSAN 8.0 to 8.0 Update 1, use the following command to manually abdicate the
DOM owner of all vSAN objects:
vsish -e set /vmkModules/vsan/dom/ownerAbdicateAll 1
Adding host back to cluster fails with the following message: A general system error occurred: Too many
outstanding requests
vSAN module unload operation can timeout while waiting for control device references. If this happens, an attempt to
move the host out of the cluster fails with the following message: Operation timed out
Any further attempts to move the host back to the cluster fail with the following message: A general system error
occurred: Too many outstanding requests
Workaround: Reboot the host before adding it back to the cluster.
Virtual machine snapshot fails after extending virtual disk size in vSAN ESA
29
VMware vSAN 8.0
This issue affects any virtual machine that has CBRC enabled in a vSAN ESA cluster. If you extend size of the VM's
virtual disks, taking a virtual machine snapshot fails.
Workaround: Perform the following steps to take a VM snapshot after you extend the size of a VM's virtual disks.
1. Power off the virtual machine and disable CBRC to all disks through API.
2. Take the virtual machine snapshot.
3. Reenable CBRC and power on the virtual machine.
Linked clone VMs migrated to vSAN ESA creates snapshot for linked clone vsanSpase disk
When migrating VMs from VMFS/NFS/vSAN OSA datastore to vSAN ESA datastore, vSAN cannot distinguish between
a snapshot vsanSparse disk and a linked clone vsanSparse disk. Since vSAN ESA supports native snapshot, a native
snapshot disk is created. If you migrate multiple VMs with moveAllDiskBackingsAndAllowSharing option, each VM
attempts to create a native snapshot of a base disk and run I/O on that object. Only the last VM can run I/O, other VMs will
fail.
Workaround: To avoid this issue, do not use moveAllDiskBackingsAndAllowSharing option when migrating linked VMs to a
vSAN ESA cluster.
hostAffinity policy option lost during upgrade
When you upgrade from vSAN 6.7 to vSAN 8.0, the vCenter Server hostaffinity option value is changed to false.
Workaround: Set the hostaffinity option back to true to continue using vSAN HostLocal policy for a normal VM.
Cannot upgrade cluster to vSAN Express Storage Architecture
You cannot upgrade or convert a cluster on vSAN Original Storage Architecture to vSAN Express Storage Architecture.
vSAN ESA is supported only on new deployments.
Workaround: None.
Encryption deep rekey not supported on vSAN ESA
vSAN Express Storage Architecture does not support encryption deep rekey in this release.
Workaround: None.
vSAN File Service not supported on vSAN ESA
vSAN Express Storage Architecture does not support vSAN File Service in this release.
Workaround: None.
Cannot change encryption settings on vSAN ESA
Encryption can only be configured vSAN ESA during cluster creation. You cannot change the settings later.
Workaround: None.
vSAN File Service does not support NFSv4 delegations
vSAN File Service does not support NFSv4 delegations in this release.
Workaround: None.
In stretched cluster, file server with no affinity cannot rebalance
In the stretched cluster vSAN File Service environment, a file server with no affinity location configured cannot be
rebalanced between Preferred ESXi hosts and Non-preferred ESXi hosts.
Workaround: Set the affinity location of the file server to Preferred or Non-Preferred by editing the file service domain
configuration.
30
VMware vSAN 8.0
Kubernetes pods with CNS volumes cannot be created, deleted, or re-scheduled during vSAN stretched cluster
partition
When a vSAN stretched cluster has a network partition between sites, an intermittent timing issue can cause volume
information to be lost from the CNS. When volume metadata is not present in the CNS, you cannot create, delete, or
re-schedule pods with CNS volumes. vSphere CSI Driver must access volume information from CNS to perform these
operations.
When the network partition is fixed, CNS volume metadata is restored, and pods with CNS volumes can be created,
deleted, or re-scheduled.
Workaround: None.
Shutdown Cluster wizard displays an error on HCI Mesh compute-ony cluster
The vSAN Shutdown Cluster wizard is designed for vSAN clusters that have a vSAN datastore and vSAN services. It does
not support HCI Mesh compute-only clusters. If you use the wizard to shutdown a compute-only cluster, it displays the
following error message:
Cannot retrieve the health service data.
Workaround: None. Do not use the vSAN Shutdown Cluster wizard on an HCI Mesh compute-only cluster.
Remediation of ESXi hosts in a vSphere Lifecycle Manager cluster with vSAN fails if vCenter services are
deployed on custom ports
If vCenter Server services are deployed on custom ports in a cluster with vSAN, vSphere DRS, and vSphere HA,
remediation of vSphere Lifecycle Manager clusters might fail. This problem is caused by a vSAN resource health check
error. ESXi hosts cannot enter maintenance mode, which leads to failing remediation tasks.
Workaround: None.
When vSAN file service is enabled, DFC-related operations such as upgrade, enabling encryption or data-
efficiency might fail
When file service is enabled, an agent VM runs on each host. The underlying vSAN object might be placed across
multiple diskgroups. When the first diskgroup gets converted, the vSAN object becomes inaccessible and the agent VM
is in an invalid state. If you try to delete the VM and redeploy a new VM, the operation fails due to the VM’s invalid state.
The VM gets unregistered but the inaccessible object still exists there. When the next diskgroup gets converted, there is
a precheck for inaccessible objects in the whole cluster. This check fails the DFC since it finds inaccessible objects of the
old agent VM.
Workaround: Manually remove the inaccessible objects.
When such failure happens, you can see the DFC task failure.
1. Identify the inaccessible objects from the failure task fault information.
2. To ensure that the objects belong to the agent VM, inspect the hostd log file and confirm that the objects belong to the
VM’s object layout.
3. Log in to the host and use the /usr/lib/vmware/osfs/bin/objtool command to remove the objects manually.
Note: To prevent this problem, disable file service before performing any DFC-related operation.
esxcli vsan cluster leave command fails to disable vSAN on an ESXi host
In some cases, the following command fails to disable vSAN on a member host: esxcli vsan cluster leave
You might see an error message similar to the following:
Failed to unmount default vSAN datastore. Unable to complete Sysinfo operation. Please see the VMKernel log file for
more details.
31
VMware vSAN 8.0
Workaround: Perform the following steps in the vSphere Client to disable vSAN on a single member host:
1. Place the host into maintenance mode.
2. Move the host out of the vSAN cluster, and into its parent data center.
vSAN service on the host is disabled automatically during the movement.
Workaround: To release the storage back to vSAN immediately, delete the file shares.
vSAN over RDMA might experience lower performance due to network congestion
RDMA requires lossless network infrastructure that is free of congestion. If your network has congestion, certain large I/O
workloads might experience lower performance than TCP.
Workaround: Address any network congestion issues following OEM best practices for RDMA.
vCenter VM crash on stretched cluster with data-in-transit encryption
vCenter VM might crash on a vSAN stretched cluster if the vCenter VM is on vSAN with data-in-transit encryption
enabled. When all hosts in one site are down and then power on again, the vCenter VM might crash after the failed site
returns to service.
Workaround: Use the following script to resolve this problem: thumbPrintRepair.py
32
VMware vSAN 8.0
With the release of vSphere Cluster Services (vCLS) in vSphere 7.0 Update 1, a set of system VMs might be placed
within the vSAN cluster. These system VMs cannot be powered-off by users. This issue can impact some vSAN
workflows, which are documented in the following article: https://kb.vmware.com/s/article/80877
Workaround: For more information about this issue, refer to this KB article: https://kb.vmware.com/s/article/80483.
vSAN File Service cannot be enabled due to an old vSAN on-disk format version
vSAN File Service cannot be enabled with the vSAN on-disk format version earlier than 11.0 (this is the on-disk format
version in vSAN 7.0).
Workaround: Upgrade the vSAN disk format version before enabling File Service.
Remediate cluster task might fail in large scale cluster due to vSAN health network test issues
Large scale clusters with more than 16 hosts, intermittent ping failures can occur during host upgrade. These failures can
interrupt host remediation in vSphere Life Cycle Manager.
Workaround: After remediation pre-check passes, silence alerts for the following vSAN health tests:
• vSAN: Basic (unicast) connectivity check
• vSAN: MTU check (ping with large packet size)
When the remediation task is complete, restore alerts for the vSAN health tests.
33
VMware vSAN 8.0
34
VMware vSAN 8.0
Reconfiguring an existing stretched cluster under a new vCenter Server causes vSAN to issue a health check
warning
When rebuilding a current stretched cluster under a new vCenter Server, the vSAN cluster health check is red. The
following message appears: vSphere cluster members match vSAN cluster members
Workaround: Use the following procedure to configure the stretched cluster.
1. Use SSH to log in to the witness host.
2. Decommission the disks on witness host. Run the following command: esxcli vsan storage remove -s "SSD
UUID"
3. Force the witness host to leave the cluster. Run the following command: esxcli vsan cluster leave
4. Reconfigure the stretched cluster from the new vCenter Server (Configure > vSAN > Fault Domains & Stretched
Cluster).
Disk format upgrade fails while vSAN resynchronizes large objects
If the vSAN cluster contains very large objects, the disk format upgrade might fail while the object is resynchronized. You
might see the following error message: Failed to convert object(s) on vSAN
vSAN cannot perform the upgrade until the object is resynchronized. You can check the status of the resynchronization
(Monitor > vSAN > Resyncing Components) to verify when the process is complete.
Workaround: Wait until no resynchronization is pending, then retry the disk format upgrade.
vSAN stretched cluster configuration lost after you disable vSAN on a cluster
If you disable vSAN on a stretched cluster, the stretched cluster configuration is not retained. The stretched cluster,
witness host, and fault domain configuration is lost.
Workaround: Reconfigure the stretched cluster parameters when you re-enable the vSAN cluster.
35
VMware vSAN 8.0
Workaround: Use RVC commands to query the state of objects in the cluster. For example: vsan.vm_object_info
Workaround: None.
HA failover does not occur after setting Traffic Type option on a vmknic to support witness traffic
36
VMware vSAN 8.0
If you set the traffic type option on a vmknic to support witness traffic, vSphere HA does not automatically discover the
new setting. You must manually disable and then re-enable HA so it can discover the vmknic. If you configure the vmknic
and the vSAN cluster first, and then enable HA on the cluster, it does discover the vmknic.
Workaround: Manually disable vSphere HA on the cluster, and then re-enable it.
37
VMware vSAN 8.0
Introduction
What's New
vSAN 8.0 introduces the following new features and enhancements:
Additional Features and Enhancements
• Enhanced network uplink latency metrics. vSAN defines more meaningful and relevant metrics catered to the
environment, whether the latencies are temporary or from an excessive workload.
• RDT level checksums. You can set checksums at the RDT layer. These new checksums can aid in debugging and
triaging.
• vSAN File Service debugging. File Service Day 0 operations have been improved for efficient validation and
troubleshooting.
• vSAN File Service over IPv6. You can create a file service domain with IPv6 network.
• vSAN File Service network reconfiguration. You can change file server IPs including the primary IP to new IPs in the
same or different subnet.
• vSphere Client Remote Plug-ins. All VMware-owned local plug-ins are transitioning to the new remote plug-in
architecture. vSAN local plug-ins have been moved to vSphere Client remote plug-ins. The local vSAN plug-ins are
deprecated in this release.
• vLCM HCL disk device. Enhancements improve vLCM’s functionality and efficiency for checking compatibility with the
desired image. It includes a check for “partNumber” and “vendor" to add coverage for more vendors.
• Reduced start time of vSAN health service. The time needed to stop vSAN health service as a part of vCenter
restart or upgrade has been reduced to 5 seconds.
• vSAN health check provides perspective to VCF LCM. This release provides only relevant vSAN health checks to
VCF in order to improve LCM resiliency in VCF.
• vSAN improves cluster NDU for VMC. New capabilities improve design and operation of a highly secure, reliable,
and operationally efficient service.
• vSAN encryption key verification. Detects invalid or corrupt keys sent from the KMS server, identifies discrepancies
between in-memory and on-disk DEKs, and alerts customers in case of discrepancies.
38
VMware vSAN 8.0
• Better handling of large component deletes. Reclaims the logical space and accounts for the physical space faster,
without causing NO_SPACE error.
• Renamed vSAN health "Check" to "Finding." This change makes the term consistent with all VMware products.
• Place vSAN in separate sandbox domain. Daemon sandboxing prevents lateral movement and provides defense in
depth. Starting with vSAN 8.0, least privilege security model is implemented, wherein any daemon that does not have
its custom sandbox domain defined, will run as a deprivileged domain. This achieves least-privilege model on an ESXi
host, with all vSAN running in their own sandbox domain with the least possible privilege.
• vSAN Proactive Insights. This mechanism enables vSAN clusters connected to VMware Analytics Cloud to identify
software and hardware anomalies proactively.
• Management and monitoring of PMEM for SAP HANA. You can manage PMEM devices within the hosts. vSAN
provides management capabilities such as health checks, performance monitoring, and space reporting for the
PMEM devices. PMEM management capabilities do not require vSAN services to be enabled. vSAN does not use
PMEM devices for caching vSAN metadata or for vSAN data services such as encryption, checksum, or dedupe and
compression. The PMEM datastore is local to each host, but can be managed from the monitor tab at the cluster level.
• Replace MD5, SHA1, and SHA2 in vSAN. SHA1 is no longer considered secure, so VMware is replacing SHA1,
MD5, and SHA2 with SHA256 across all VMware products, including vSAN.
• IL6 compliance. vSAN 8.0 is IL6 compliant.
39
VMware vSAN 8.0
40
VMware vSAN 8.0
Disk format version 1.0 does not have performance and snapshot enhancements, and it lacks support for advanced
features including checksum, deduplication and compression, and encryption. For more information about vSAN disk
format version, see KB 2148493.
Upgrading the On-disk Format for Hosts with Limited Capacity
During an upgrade of the vSAN on-disk format from version 1.0 or 2.0, a disk group evacuation is performed. The disk
group is removed and upgraded to on-disk format version 17.0, and the disk group is added back to the cluster. For two-
node or three-node clusters, or clusters without enough capacity to evacuate each disk group, select Allow Reduced
Redundancy from the vSphere Client. You also can use the following RVC command to upgrade the on-disk format:
vsan.ondisk_upgrade --allow-reduced-redundancy
When you allow reduced redundancy, your VMs are unprotected for the duration of the upgrade, because this method
does not evacuate data to the other hosts in the cluster. It removes each disk group, upgrades the on-disk format, and
adds the disk group back to the cluster. All objects remain available, but with reduced redundancy.
If you enable deduplication and compression during the upgrade to vSAN 8.0, you can select Allow Reduced
Redundancy from the vSphere Client.
Limitations
For information about maximum configuration limits for the vSAN 8.0 release, see the Configuration Maximums
documentation.
Resolved Issues
RemoveFileShare task failure may cause vSAN File Services server failover
RemoveFileShare task for the NFS share may fail on the VCenter Server even though the share is deleted. This happens
because the NFS server fails while removing the export. This does not cause any problems in the overall workflow as the
share gets successfully deleted.
When the NFS server fails, it triggers vSAN File Services server failover. Since NFS server and SMB server fails
over together, if there are any SMB shares exported from the same vSAN File Services server it causes SMB mount
disruptions. SMB mount disruption due to server failover is a known behavior as vSAN does not support transparent
failover for SMB servers.
Workaround: None.
vSAN Health cannot find VUM with proxy configured
When a proxy is configured for vSAN, the vsan-health service falsely reported that VMware Update Manager (VUM) is
disabled or not installed.
This issue is fixed in this release.
Known Issues
41
VMware vSAN 8.0
Workaround: None.
Cannot change encryption settings on vSAN ESA
Encryption can only be configured vSAN ESA during cluster creation. You cannot change the settings later.
Workaround: None.
vSAN File Service not supported on vSAN ESA
vSAN Express Storage Architecture does not support vSAN File Service in this release.
Workaround: None.
Encryption deep rekey not supported on vSAN ESA
vSAN Express Storage Architecture does not support encryption deep rekey in this release.
Workaround: None.
Cannot upgrade cluster to vSAN Express Storage Architecture
You cannot upgrade or convert a cluster on vSAN Original Storage Architecture to vSAN Express Storage Architecture.
vSAN ESA is supported only on new deployments.
Workaround: None.
hostAffinity policy option lost during upgrade
When you upgrade from vSAN 6.7 to vSAN 8.0, the vCenter Server hostaffinity option value is changed to false.
Workaround: Set the hostaffinity option back to true to continue using vSAN HostLocal policy for a normal VM.
Kubernetes pods with CNS volumes cannot be created, deleted, or re-scheduled during vSAN stretched cluster
partition
When a vSAN stretched cluster has a network partition between sites, an intermittent timing issue can cause volume
information to be lost from the CNS. When volume metadata is not present in the CNS, you cannot create, delete, or
re-schedule pods with CNS volumes. vSphere CSI Driver must access volume information from CNS to perform these
operations.
When the network partition is fixed, CNS volume metadata is restored, and pods with CNS volumes can be created,
deleted, or re-scheduled.
Workaround: None.
Shutdown Cluster wizard displays an error on HCI Mesh compute-ony cluster
The vSAN Shutdown Cluster wizard is designed for vSAN clusters that have a vSAN datastore and vSAN services. It does
not support HCI Mesh compute-only clusters. If you use the wizard to shutdown a compute-only cluster, it displays the
following error message:
Cannot retrieve the health service data.
Workaround: None. Do not use the vSAN Shutdown Cluster wizard on an HCI Mesh compute-only cluster.
Remediation of ESXi hosts in a vSphere Lifecycle Manager cluster with vSAN fails if vCenter services are
deployed on custom ports
If vCenter Server services are deployed on custom ports in a cluster with vSAN, vSphere DRS, and vSphere HA,
remediation of vSphere Lifecycle Manager clusters might fail. This problem is caused by a vSAN resource health check
error. ESXi hosts cannot enter maintenance mode, which leads to failing remediation tasks.
Workaround: None.
42
VMware vSAN 8.0
When vSAN file service is enabled, DFC-related operations such as upgrade, enabling encryption or data-
efficiency might fail
When file service is enabled, an agent VM runs on each host. The underlying vSAN object might be placed across
multiple diskgroups. When the first diskgroup gets converted, the vSAN object becomes inaccessible and the agent VM
is in an invalid state. If you try to delete the VM and redeploy a new VM, the operation fails due to the VM’s invalid state.
The VM gets unregistered but the inaccessible object still exists there. When the next diskgroup gets converted, there is
a precheck for inaccessible objects in the whole cluster. This check fails the DFC since it finds inaccessible objects of the
old agent VM.
Workaround: Manually remove the inaccessible objects.
When such failure happens, you can see the DFC task failure.
1. Identify the inaccessible objects from the failure task fault information.
2. To ensure that the objects belong to the agent VM, inspect the hostd log file and confirm that the objects belong to the
VM’s object layout.
3. Log in to the host and use the /usr/lib/vmware/osfs/bin/objtool command to remove the objects manually.
Note: To prevent this problem, disable file service before performing any DFC-related operation.
esxcli vsan cluster leave command fails to disable vSAN on an ESXi host
In some cases, the following command fails to disable vSAN on a member host: esxcli vsan cluster leave
You might see an error message similar to the following:
Failed to unmount default vSAN datastore. Unable to complete Sysinfo operation. Please see the VMKernel log file for
more details.
Workaround: Perform the following steps in the vSphere Client to disable vSAN on a single member host:
1. Place the host into maintenance mode.
2. Move the host out of the vSAN cluster, and into its parent data center.
vSAN service on the host is disabled automatically during the movement.
Workaround: To release the storage back to vSAN immediately, delete the file shares.
vSAN over RDMA might experience lower performance due to network congestion
RDMA requires lossless network infrastructure that is free of congestion. If your network has congestion, certain large I/O
workloads might experience lower performance than TCP.
Workaround: Address any network congestion issues following OEM best practices for RDMA.
43
VMware vSAN 8.0
44
VMware vSAN 8.0
Remediate cluster task might fail in large scale cluster due to vSAN health network test issues
Large scale clusters with more than 16 hosts, intermittent ping failures can occur during host upgrade. These failures can
interrupt host remediation in vSphere Life Cycle Manager.
Workaround: After remediation pre-check passes, silence alerts for the following vSAN health tests:
• vSAN: Basic (unicast) connectivity check
• vSAN: MTU check (ping with large packet size)
When the remediation task is complete, restore alerts for the vSAN health tests.
45
VMware vSAN 8.0
Workaround: None.
Unable to remove a dedupe enabled disk group after a capacity disk enters PDL state
When a capacity disk in a dedupe-enabled disk group is removed, or its unique ID changes, or when the device
experiences an unrecoverable hardware error, it enters Permanent Device Loss (PDL) state. If you try to remove the disk
group, you might see an error message informing you that the action cannot be completed.
Workaround: Whenever a capacity disk is removed, or its unique ID changes, or when the device experiences an
unrecoverable hardware error, wait for a few minutes before trying to remove the disk group.
vSAN health indicates non-availability related incompliance with failed pending policy
A policy change request leaves the object health status of vSAN in a non-availability related incompliance state. This is
because there might be other scheduled work that is utilizing the requested resources. However, vSAN reschedules this
policy request automatically as resources become available.
Workaround: The vSAN period scan fixes this issue automatically in most cases. However, other work in progress might
use up available resources even after the policy change was accepted but not applied. You can add more capacity if the
capacity reporting displays a high value.
In deduplication clusters, reactive rebalancing might not happen when the disks show more than 80% full
In deduplication clusters, when the disks display more than 80% full on the dashboard, the reactive rebalancing might not
start as expected. This is because in deduplication clusters, pending writes and deletes are also considered for calculating
the free capacity.
Workaround: None.
TRIM/UNMAP commands from Guest OS fail
If the Guest OS attempts to perform space reclamation during online snapshot consolidation, the TRIM/UNMAP
commands fail. This failure keeps space from being reclaimed.
Workaround: Try to reclaim the space after the online snapshot operation is complete. If subsequent TRIM/UNMAP
operations fail, remount the disk.
Space reclamation from SCSI TRIM/UNMAP is lost when online snapshot consolidation is performed
Space reclamation achieved from SCSI TRIM/UNMAP commands is lost when you perform online snapshot consolidation.
Offline snapshot consolidation does not affect SCSI unmap operation.
Workaround: Reclaim the space after online snapshot consolidation is complete.
46
VMware vSAN 8.0
When rebuilding a current stretched cluster under a new vCenter Server, the vSAN cluster health check is red. The
following message appears: vSphere cluster members match vSAN cluster members
Workaround: Use the following procedure to configure the stretched cluster.
1. Use SSH to log in to the witness host.
2. Decommission the disks on witness host. Run the following command: esxcli vsan storage remove -s "SSD
UUID"
3. Force the witness host to leave the cluster. Run the following command: esxcli vsan cluster leave
4. Reconfigure the stretched cluster from the new vCenter Server (Configure > vSAN > Fault Domains & Stretched
Cluster).
Disk format upgrade fails while vSAN resynchronizes large objects
If the vSAN cluster contains very large objects, the disk format upgrade might fail while the object is resynchronized. You
might see the following error message: Failed to convert object(s) on vSAN
vSAN cannot perform the upgrade until the object is resynchronized. You can check the status of the resynchronization
(Monitor > vSAN > Resyncing Components) to verify when the process is complete.
Workaround: Wait until no resynchronization is pending, then retry the disk format upgrade.
vSAN stretched cluster configuration lost after you disable vSAN on a cluster
If you disable vSAN on a stretched cluster, the stretched cluster configuration is not retained. The stretched cluster,
witness host, and fault domain configuration is lost.
Workaround: Reconfigure the stretched cluster parameters when you re-enable the vSAN cluster.
47
VMware vSAN 8.0
After you perform a force repair, some objects might not be repaired because the ownership of the objects was transferred
to a different node during the process. The force repair might be delayed for those objects.
Workaround: Attempt the force repair operation after all other objects are repaired and resynchronized. You can wait until
vSAN repairs the objects.
When you move a host from one encrypted cluster to another, and then back to the original cluster, the task fails
When you move a host from an encrypted vSAN cluster to another encrypted vSAN cluster, then move the host back to
the original encrypted cluster, the task might fail. You might see the following message: A general system error
occurred: Invalid fault . This error occurs because vSAN cannot re-encrypt data on the host using the original
encryption key. After a short time, vCenter Server restores the original key on the host, and all unmounted disks in the
vSAN cluster are mounted.
Workaround: Reboot the host and wait for all disks to get mounted.
Stretched cluster imbalance after a site recovers
When you recover a failed site in a stretched cluster, sometimes hosts in the failed site are brought back sequentially over
a long period of time. vSAN might overuse some hosts when it begins repairing the absent components.
Workaround: Recover all of the hosts in a failed site together within a short time window.
VM operations fail due to HA issue with stretched clusters
Under certain failure scenarios in stretched clusters, certain VM operations such as vMotions or powering on a VM might
be impacted. These failures scenarios include a partial or a complete site failure, or the failure of the high speed network
between the sites. This problem is caused by the dependency on VMware HA being available for normal operation of
stretched cluster sites.
Workaround: Disable vSphere HA before performing vMotion, VM creation, or powering on VMs. Then re-enable vSphere
HA.
Cannot perform deep rekey if a disk group is unmounted
Before vSAN performs a deep rekey, it performs a shallow rekey. The shallow rekey fails if an unmounted disk group is
present. The deep rekey process cannot begin.
Workaround: Remount or remove the unmounted disk group.
Log entries state that firewall configuration has changed
A new firewall entry appears in the security profile when vSAN encryption is enabled: vsanEncryption. This rule controls
how hosts communicate directly to the KMS. When it is triggered, log entries are added to /var/log/vobd.log . You
might see the following messages:
Firewall configuration has changed. Operation 'addIP4' for rule set vsanEncryption
succeeded.
Firewall configuration has changed. Operation 'removeIP4' for rule set vsanEncryption
succeeded.
These messages can be ignored.
Workaround: None.
HA failover does not occur after setting Traffic Type option on a vmknic to support witness traffic
If you set the traffic type option on a vmknic to support witness traffic, vSphere HA does not automatically discover the
new setting. You must manually disable and then re-enable HA so it can discover the vmknic. If you configure the vmknic
and the vSAN cluster first, and then enable HA on the cluster, it does discover the vmknic.
Workaround: Manually disable vSphere HA on the cluster, and then re-enable it.
48
VMware vSAN 8.0
49
VMware vSAN 8.0
Intended Audience
This manual is intended for anyone who wants to design and deploy avSAN cluster in a VMware vSphere environment.
The information in this manual is written for experienced system administrators who are familiar with virtual machine
technology and virtual datacenter operations. This manual assumes familiarity with VMware vSphere, including VMware
ESXi, vCenter Server, and the vSphere Client.
For more information about vSAN features and how to configure avSAN cluster, see Administering VMware vSAN.
For more information about monitoring avSAN cluster and fixing problems, see the vSAN Monitoring and Troubleshooting
Guide.
Updated Information
This document is updated with each release of the product or when necessary.
This table provides the update history of vSAN Planning and Deployment.
Revision Description
What Is vSAN
VMware vSAN is a distributed layer of software that runs natively as a part of the ESXi hypervisor.
vSAN aggregates local or direct-attached capacity devices of a host cluster and creates a single storage pool shared
across all hosts in the vSAN cluster. While supporting VMware features that require shared storage, such as HA, vMotion,
and DRS, vSAN eliminates the need for external shared storage and simplifies storage configuration and virtual machine
provisioning activities.
50
VMware vSAN 8.0
vSAN Concepts
VMware vSAN uses a software-defined approach that creates shared storage for virtual machines.
It virtualizes the local physical storage resources of ESXi hosts and turns them into pools of storage that can be
divided and assigned to virtual machines and applications according to their quality-of-service requirements. vSAN is
implemented directly in the ESXi hypervisor.
You can configure vSAN to work as either a hybrid or all-flash cluster. In hybrid clusters, flash devices are used for the
cache layer and magnetic disks are used for the storage capacity layer. In all-flash clusters, flash devices are used for
both cache and capacity.
You can activate vSAN on existing host clusters, or when you create a new cluster. vSAN aggregates all local capacity
devices into a single datastore shared by all hosts in the vSAN cluster. You can expand the datastore by adding capacity
devices or hosts with capacity devices to the cluster. vSAN works best when all ESXi hosts in the cluster share similar or
identical configurations across all cluster members, including similar or identical storage configurations. This consistent
configuration balances virtual machine storage components across all devices and hosts in the cluster. Hosts without any
local devices also can participate and run their virtual machines on the vSAN datastore.
In vSAN Original Storage Architecture (OSA), each host that contributes storage devices to the vSAN datastore must
provide at least one device for flash cache and at least one device for capacity. The devices on the contributing host
form one or more disk groups. Each disk group contains one flash cache device, and one or multiple capacity devices for
persistent storage. Each host can be configured to use multiple disk groups.
In vSAN Express Storage Architecture (ESA), all storage devices claimed by vSAN contribute to capacity and
performance. Each host's storage devices claimed by vSAN form a storage pool. The storage pool represents the amount
of caching and capacity provided by the host to the vSAN datastore.
For best practices, capacity considerations, and general recommendations about designing and sizing a vSAN cluster,
see the VMware vSAN Design and Sizing Guide.
Characteristics of vSAN
The following characteristics apply to vSAN, its clusters, and datastores.
vSAN includes numerous features to add resiliency and efficiency to your data computing and storage environment.
Shared storage support vSAN supports VMware features that require shared storage, such as HA,
vMotion, and DRS. For example, if a host becomes overloaded, DRS can
migrate virtual machines to other hosts in the cluster.
On-disk format vSAN on-disk virtual file format provides highly scalable snapshot and
clone management support per vSAN cluster. For information about
the number of virtual machine snapshots and clones supported per
vSAN cluster, refer to the vSphere Configuration Maximumshttps://
configmax.esp.vmware.com/home.
All-flash and hybrid configurations vSAN can be configured for all-flash or hybrid cluster.
Fault domains vSAN supports configuring fault domains to protect hosts from rack or
chassis failures when the vSAN cluster spans across multiple racks or
blade server chassis in a data center.
File service vSAN file service enables you to create file shares in the vSAN datastore
that client workstations or VMs can access.
51
VMware vSAN 8.0
iSCSI target service vSAN iSCSI target service enables hosts and physical workloads that
reside outside the vSAN cluster to access the vSAN datastore.
vSAN Stretched cluster and Two node vSAN cluster vSAN supports stretched clusters that span across two geographic
locations.
Support for Windows Server Failover Clusters (WSFC) vSAN 6.7 Update 3 and later releases support SCSI-3 Persistent
Reservations (SCSI3-PR) on a virtual disk level required by Windows
Server Failover Cluster (WSFC) to arbitrate an access to a shared disk
between nodes. Support of SCSI-3 PRs enables configuration of WSFC
with a disk resource shared between VMs natively on vSAN datastores.
Currently the following configurations are supported:
• Up to 6 application nodes per cluster.
• Up to 64 shared virtual disks per node.
NOTE
Microsoft SQL Server 2012 or later running on Microsoft
Windows Server 2012 or later has been qualified on vSAN.
vSAN health service vSAN health service includes preconfigured health check tests to monitor,
troubleshoot, diagnose the cause of cluster component problems, and
identify any potential risk.
vSAN performance service vSAN performance service includes statistical charts used to monitor IOPS,
throughput, latency, and congestion. You can monitor performance of a
vSAN cluster, host, disk group, disk, and VMs.
Integration with vSphere storage features vSAN integrates with vSphere data management features traditionally used
with VMFS and NFS storage. These features include snapshots, linked
clones, and vSphere Replication.
Virtual Machine Storage Policies vSAN works with VM storage policies to support a VM-centric approach to
storage management.
If you do not assign a storage policy to the virtual machine during
deployment, the vSAN Default Storage Policy is automatically assigned to
the VM.
®
Rapid provisioning vSAN enables rapid provisioning of storage in the vCenter Server during
virtual machine creation and deployment operations.
Deduplication and compression vSAN performs block-level deduplication and compression to save
storage space. When you enable deduplication and compression on a
vSAN all-flash cluster, redundant data within each disk group is reduced.
Deduplication and compression is a cluster-wide setting, but the functions
are applied on a disk group basis. Compression-only vSAN is applied on a
per-disk basis.
Data at rest encryption vSAN provides data at rest encryption. Data is encrypted after all other
processing, such as deduplication, is performed. Data at rest encryption
protects data on storage devices, in case a device is removed from the
cluster.
Data in transit encryption vSAN can encrypt data in transit across hosts in the cluster. When you
enable data-in-transit encryption, vSAN encrypts all data and metadata
traffic between hosts.
52
VMware vSAN 8.0
SDK support The VMware vSAN SDK is an extension of the VMware vSphere
Management SDK. It includes documentation, libraries and code examples
that help developers automate installation, configuration, monitoring, and
troubleshooting of vSAN.
Consumed Capacity
Consumed capacity is the amount of physical capacity consumed by one or more virtual machines at any point. Many
factors determine consumed capacity, including the consumed size of your .vmdk files, protection replicas, and so on.
When calculating for cache sizing, do not consider the capacity used for protection replicas.
Object-Based Storage
vSAN stores and manages data in the form of flexible data containers called objects. An object is a logical volume that
has its data and metadata distributed across the cluster. For example, every .vmdk is an object, as is every snapshot.
When you provision a virtual machine on a vSAN datastore, vSAN creates a set of objects comprised of multiple
components for each virtual disk. It also creates the VM home namespace, which is a container object that stores all
metadata files of your virtual machine. Based on the assigned virtual machine storage policy, vSAN provisions and
manages each object individually, which might also involve creating a RAID configuration for every object.
NOTE
If vSAN Express Storage Architecture is enabled, every snapshot is not a new object. A base .vmdk and its
snapshots are contained in one vSAN object. Additionally, in vSAN ESA, digest is backed by vSAN objects.
53
VMware vSAN 8.0
When vSAN creates an object for a virtual disk and determines how to distribute the object in the cluster, it considers the
following factors:
• vSAN verifies that the virtual disk requirements are applied according to the specified virtual machine storage policy
settings.
• vSAN verifies that the correct cluster resources are used at the time of provisioning. For example, based on the
protection policy, vSAN determines how many replicas to create. The performance policy determines the amount of
flash read cache allocated for each replica and how many stripes to create for each replica and where to place them in
the cluster.
• vSAN continually monitors and reports the policy compliance status of the virtual disk. If you find any noncompliant
policy status, you must troubleshoot and resolve the underlying problem.
NOTE
When required, you can edit VM storage policy settings. Changing the storage policy settings does not affect
virtual machine access. vSAN actively throttles the storage and network resources used for reconfiguration
to minimize the impact of object reconfiguration to normal workloads. When you change VM storage policy
settings, vSAN might initiate an object recreation process and subsequent resynchronization. See vSAN
Monitoring and Troubleshooting.
• vSAN verifies that the required protection components, such as mirrors and witnesses, are placed on separate hosts
or fault domains. For example, to rebuild components during a failure, vSAN looks for ESXi hosts that satisfy the
placement rules where protection components of virtual machine objects must be placed on two different hosts, or
across fault domains.
vSAN Datastore
After you enable vSAN on a cluster, a single vSAN datastore is created. It appears as another type of datastore in the list
of datastores that might be available, including Virtual Volume, VMFS, and NFS. A single vSAN datastore can provide
different service levels for each virtual machine or each virtual disk. In vCenter Server®, storage characteristics of the
vSAN datastore appear as a set of capabilities. You can reference these capabilities when defining a storage policy for
virtual machines. When you later deploy virtual machines, vSAN uses this policy to place virtual machines in the optimal
manner based on the requirements of each virtual machine. For general information about using storage policies, see the
vSphere Storage documentation.
A vSAN datastore has specific characteristics to consider.
• vSAN provides a single vSAN datastore accessible to all hosts in the cluster, whether or not they contribute storage to
the cluster. Each host can also mount any other datastores, including Virtual Volumes, VMFS, or NFS.
• You can use Storage vMotion to move virtual machines between vSAN datastores, NFS datastores, and VMFS
datastores.
• Only magnetic disks and flash devices used for capacity can contribute to the datastore capacity. The devices used for
flash cache are not counted as part of the datastore.
54
VMware vSAN 8.0
VMDK
A virtual machine disk or .vmdk file that stores the contents of the virtual machine's hard disk drive.
VM Swap Object
Created when a virtual machine is powered on.
Snapshot Delta VMDKs
Created when virtual machine snapshots are taken. Such delta disks are not created for vSAN Express Storage Architecture.
Memory object
Created when the snapshot memory option is selected when creating or suspending a virtual machine.
Witness
A witness is a component that contains only metadata and does not contain any actual application data. It serves as
a tiebreaker when a decision must be made regarding the availability of the surviving datastore components, after a
potential failure. A witness consumes approximately 2 MB of space for metadata on the vSAN datastore when using on-
disk format 1.0, and 4 MB for on-disk format version 2.0 and later.
vSAN maintains a quorum by using an asymmetrical voting system where each component might have more than one
vote to decide the availability of objects. Greater than 50 percent of the votes that make up a VM’s storage object must
be accessible at all times for the object to be considered available. When 50 percent or fewer votes are accessible to all
hosts, the object is no longer accessible to the vSAN datastore. Inaccessible objects can impact the availability of the
associated VM.
55
VMware vSAN 8.0
vSphere PowerCLI
VMware vSphere PowerCLI adds command-line scripting support for vSAN, to help you automate configuration and
management tasks. vSphere PowerCLI provides a Windows PowerShell interface to the vSphere API. PowerCLI includes
cmdlets for administering vSAN components. For information about using vSphere PowerCLI, see vSphere PowerCLI
Documentation.
56
VMware vSAN 8.0
Depending on your requirement, you can deploy vSAN in the following ways.
vSAN ReadyNode
The vSAN ReadyNode is a preconfigured solution of the vSAN software provided by VMware partners, such as Cisco,
Dell, HPE, Fujitsu, IBM, and Supermicro. This solution includes validated server configuration in a tested, certified
hardware form factor for vSAN deployment that is recommended by the server OEM and VMware. For information about
the vSAN ReadyNode solution for a specific partner, visit the VMware Partner website.
57
VMware vSAN 8.0
58
VMware vSAN 8.0
vSphere HA
You can enable vSphere HA and vSAN on the same cluster. As with traditional datastores, vSphere HA provides the same
level of protection for virtual machines on vSAN datastores. This level of protection imposes specific restrictions when
vSphere HA and vSAN interact. For specific considerations about integrating vSphere HA and vSAN, see Using vSAN
and vSphere HA.
59
VMware vSAN 8.0
Limitations of vSAN
This topic discusses the limitations of vSAN.
When working with vSAN, consider the following limitations:
• vSAN does not support hosts participating in multiple vSAN clusters. However, a vSAN host can access other external
storage resources that are shared across clusters.
• vSAN does not support vSphere DPM and Storage I/O Control.
• vSAN does not support SE Sparse disks.
• vSAN does not support RDM, VMFS, diagnostic partition, and other device access features.
Cache • One SAS or SATA solid-state disk (SSD) or PCIe flash device.
• Before calculating the Failures to tolerate, check the size of the flash
caching device in each disk group. For hybrid cluster, it must provide at
least 10 percent of the anticipated storage consumed on the capacity
devices, not including replicas such as mirrors.
• vSphere Flash Read Cache must not use any of the flash devices
reserved for vSAN cache.
• The cache flash devices must not be formatted with VMFS or another file
system.
Capacity • Hybrid group configuration must have at least one SAS or NL-SAS
magnetic disk.
• All-flash disk group configuration must have at least one SAS, or SATA
solid-state disk (SSD), or PCIe flash device.
60
VMware vSAN 8.0
Storage controllers One SAS or SATA host bus adapter (HBA), or a RAID controller that is in
passthrough mode or RAID 0 mode.
To avoid issues, consider these points when the same storage controller is
backing both vSAN and non-vSAN disks:
Do not mix the controller mode for vSAN and non-vSAN disks to avoid
handling the disks inconsistently, which can negatively impact vSAN
operation. If the vSAN disks are in RAID mode, the non-vSAN disks must
also be in RAID mode.
When you use non-vSAN disks for VMFS, use the VMFS datastore only for
scratch, logging, and core dumps.
Do not run virtual machines from a disk or RAID group that shares its
controller with vSAN disks or RAID groups.
Do not passthrough non-vSAN disks to virtual machine guests as Raw
Device Mappings (RDMs).
To learn about controller supported features, such as passthrough and
RAID, refer to the vSAN HCL: https://www.vmware.com/resources/
compatibility/search.php?deviceCategory=vsan
Cache and capacity Each storage pool must have at least one NVMe TLC device.
Host Memory
The memory requirements for vSAN Original Storage Architecture depend on the number of disk groups and devices
that the ESXi hypervisor must manage. For more information, see the VMware knowledge base article at https://
kb.vmware.com/s/article/2113954.
vSAN Express Storage Architecture requires at least 128 GB host memory. The memory needed for your environment
depends on the number of devices in the host's storage pool.
61
VMware vSAN 8.0
Host Bandwidth Each host must have minimum bandwidth dedicated to vSAN.
• vSAN OSA: Dedicated 1 Gbps for hybrid configurations, dedicated or
shared 10 Gbps for all-flash configurations
• vSAN ESA: Dedicated or shared 10 Gbps
For information about networking considerations in vSAN, see Designing the
vSAN Network.
Connection between hosts Each host in the vSAN cluster, regardless of whether it contributes capacity,
must have a VMkernel network adapter for vSAN traffic. See Set Up a
VMkernel Network for vSAN.
Host network All hosts in your vSAN cluster must be connected to a vSAN Layer 2 or Layer 3
network.
IPv4 and IPv6 support The vSAN network supports both IPv4 and IPv6.
Network latency • Maximum of 1 ms RTT for single site (non-stretched) vSAN clusters
between all hosts in the cluster
• Maximum of 5 ms RTT between the two main sites for vSAN stretched
clusters
• Maximum of 200 ms RTT from a main site to the vSAN witness host
License Requirements
vSAN clusters are licensed differently with the per TiB, per CPU, and per Core licensing model.
In a vSphere environment converted to VMware Cloud-connection based vSphere+ subscription, you can continue to use
vSAN CPU license keys. For more information, see the VMware vSphere+ documentation.
62
VMware vSAN 8.0
The license use of the vSAN is recalculated and updated in the following cases:
• If you assign a new license to the vSAN cluster
• If you add a new host to the vSAN cluster
• If a host is removed from the cluster
• If the total number of TiBs in a cluster changes
You need to purchase a vSAN add-on license if you need additional capacity. For capacity larger than the total entitled
vSAN capacity in tebibytes, you can purchase additional vSAN capacity. When you purchase additional capacity, you
receive a vSAN capacity license. You can combine multiple license keys and apply the resulting license key to vSAN
clusters.
63
VMware vSAN 8.0
To calculate the capacity that you need for your vSAN environment, you need the total number of licensed CPU cores for
each CPU on all the ESXi hosts in your environment. For example, consider a vSAN cluster with 3 ESXi hosts, 1 CPU
per host, and each host has 8 physical cores per CPU. You can use up to 0.25 TiB of included vSAN storage per vSAN
host physical core. You must purchase a vSphere Foundation license with the subscription capacity of 16 cores per CPU
because it is the required minimum license capacity. With a total of 48 (3 * 1 * 16) physical cores per CPU, you receive 12
TiBs (0.25 TiB * total cores in the vSAN cluster) capacity. Similarly, With a total of 128 (4 * 2* 16) physical cores per CPU,
you receive 32 TiBs capacity.
You need to purchase an add-on license if you need additional capacity. The vSAN clusters with more than 0.25 TiB
core of storage requires a vSAN add-on license for the entire storage capacity of the cluster. For more information about
calculating the license capacity that you need for your environment, see the VMware knowledge base article at https://
kb.vmware.com/s/article/95927.
64
VMware vSAN 8.0
For example, if you have 1 ESXi host with 1 CPU, and 8 CPU cores per CPU, you must purchase the subscription
capacity of 16 cores per CPU because it is the minimum license capacity.
Number of ESXi Hosts Number of CPUs Cores per CPU Number of Core Licenses
1 1 8 16
2 2 8 64
2 2 16 64
For more information about calculating the number of licenses you need for your environment, see the VMware
knowledge base article at https://kb.vmware.com/s/article/95927.
Raw Capacity
Use this formula to determine the raw capacity of a vSAN datastore. Multiply the total number of disk groups in the cluster
by the size of the capacity devices in those disk groups. Subtract the overhead required by the vSAN on-disk format.
Failures to Tolerate
When you plan the capacity of the vSAN datastore, not including the number of virtual machines and the size of their
VMDK files, you must consider the Failures to tolerate of the virtual machine storage policies for the cluster.
65
VMware vSAN 8.0
The Failures to tolerate has an important role when you plan and size storage capacity for vSAN. Based on the
availability requirements of a virtual machine, the setting might result in doubled consumption or more, compared with the
consumption of a virtual machine and its individual devices.
For example, if the Failures to tolerate is set to 1 failure - RAID-1 (Mirroring), virtual machines can use about 50
percent of the raw capacity. If the FTT is set to 2, the usable capacity is about 33 percent. If the FTT is set to 3, the usable
capacity is about 25 percent.
But if the Failures to tolerate is set to 1 failure - RAID-5 (Erasure Coding), virtual machines can use about 75 percent
of the raw capacity. If the FTT is set to 2 failures - RAID-6 (Erasure Coding), the usable capacity is about 67 percent.
For more information about RAID 5/6, see Administering VMware vSAN.
For information about the attributes in a vSAN storage policy, see Administering VMware vSAN.
66
VMware vSAN 8.0
The space that is required might be different. Its size depends on how often the virtual machine changes data and how
long a snapshot is attached to the virtual machine.
• Swap files. In vSAN 6.7 and later, virtual machine swap files inherit the storage policy of the VM Namespace.
67
VMware vSAN 8.0
All-flash and hybrid configurations • A higher cache-to-capacity ratio eases future capacity growth. Oversizing
cache enables you to add more capacity to an existing disk group without
the need to increase the size of the cache.
• Flash caching devices must have high write endurance.
• Replacing a flash caching device is more complicated than replacing a
capacity device because such an operation impacts the entire disk group.
• If you add more flash devices to increase the size of the cache, you must
create more disk groups. The ratio between flash cache devices and disk
groups is always 1:1.
A configuration of multiple disk groups provides the following advantages:
– Reduced risk of failure. If a single caching device fails, fewer capacity
devices are affected.
– Potentially improved performance if you deploy multiple disk groups that
contain smaller flash caching devices.
However, when you configure multiple disk groups, the memory
consumption of the hosts increases.
All-flash configurations In all-flash configurations, vSAN uses the cache layer for write caching only.
The write cache must be able to handle high write activities. This approach
extends the life of capacity flash that might be less expensive and might have
lower write endurance.
Hybrid configurations The flash caching device must provide at least 10 percent of the anticipated
storage that virtual machines are expected to consume, not including replicas
such as mirrors. The Primary level of failures to tolerate attribute from the
VM storage policy does not impact the size of the cache.
If the read cache reservation is configured in the active VM storage policy, the
hosts in the vSAN cluster must have sufficient cache to satisfy the reservation
during a post-failure rebuild or maintenance operation.
If the available read cache is not sufficient to satisfy the reservation, the rebuild
or maintenance operation fails. Use read cache reservation only if you must
meet a specific, known performance requirement for a particular workload.
The use of snapshots consumes cache resources. If you plan to use several
snapshots, consider dedicating more cache than the conventional 10 percent
cache-to-consumed-capacity ratio.
68
VMware vSAN 8.0
• Write endurance. The write endurance of the SSD devices must meet the requirements for capacity or for cache in all-
flash configurations, and for cache in hybrid configurations.
For information about the write endurance requirements for all-flash and hybrid configurations, see the VMware
Design and Sizing Guide. For information about the write endurance class of SSD devices, see the section of the
VMware Compatibility Guide.
• Cost. PCIe devices generally have higher cost than SSD devices.
69
VMware vSAN 8.0
70
VMware vSAN 8.0
Host Networking
Provide more bandwidth for vSAN traffic to improve performance.
• vSAN Original Storage Architecture
– If you plan to use hosts that have 1-GbE adapters, dedicate adapters for vSAN only. For all-flash configurations,
plan hosts that have dedicated or shared 10-GbE adapters.
– If you plan to use 10-GbE adapters, they can be shared with other traffic types for both hybrid and all-flash
configurations.
• vSAN Express Storage Architecture
– Plan to use hosts that have dedicated or shared 25-GbE adapters or better.
– Network adapters can be shared with other traffic types.
• If a network adapter is shared with other traffic types, use a vSphere Distributed Switch to isolate vSAN traffic by using
Network I/O Control and VLANs.
• Create a team of physical adapters to provide redundancy for vSAN traffic.
Drive Bays
For easy maintenance, consider hosts whose drive bays and PCIe slots are at the front of the server body.
71
VMware vSAN 8.0
72
VMware vSAN 8.0
If you plan to configure a NIC team for availability, consider these failover configurations.
vSAN supports IP-hash load balancing, but cannot guarantee improvement in performance for all configurations. You can
benefit from IP hash when vSAN is among its many consumers. In this case, IP hash performs load balancing. If vSAN
is the only consumer, you might observe no improvement. This behavior specifically applies to 1-GbE environments. For
example, if you use four 1-GbE physical adapters with IP hash for vSAN, you might not be able to use more than 1 Gbps.
This behavior also applies to all NIC teaming policies that VMware supports.
vSAN does not support multiple VMkernel adapters on the same subnet. You can use different VMkernel adapters on
different subnets, such as another VLAN or separate physical fabric. Providing availability by using several VMkernel
adapters has configuration costs that involve vSphere and the network infrastructure. You can increase network
availability by teaming physical network adapters.
Using RDMA
vSAN 7.0 Update 2 and later releases can use Remote Direct Memory Access (RDMA). RDMA typically has lower CPU
utilization and less I/O latency. If your hosts support the RoCE v2 protocol, you can enable RDMA through the vSAN
network service in vSphere Client.
Consider the following guidelines when designing vSAN over RDMA:
• Each vSAN host must have a vSAN certified RDMA-capable NIC, as listed in the vSAN section of the VMware
Compatibility Guide. Use only the same model network adapters from the same vendor on each end of the connection.
Configure the DCBx mode to IEEE.
• All hosts must support RDMA. If any host loses RDMA support, the entire vSAN cluster switches to TCP.
• The network must be lossless. Configure network switches to use Data Center Bridging with Priority Flow Control.
Configure a lossless traffic class for vSAN traffic marked at priority level 3.
• vSAN with RDMA does not support LACP or IP-hash-based NIC teaming. vSAN with RDMA does support NIC failover.
• All hosts must be on the same subnet. vSAN with RDMA supports up to 32 hosts.
73
VMware vSAN 8.0
Table 7: Example Network I/O Control Configuration for a Physical Adapter That Handles vSAN
vSAN 1 100
vSphere vMotion 0.5 70
Virtual machine 0.5 30
If the network adapter becomes saturated, Network I/O Control allocates 5 Gbps to vSAN on the physical adapter.
For information about using vSphere Network I/O Control to configure bandwidth allocation for vSAN traffic, see the
vSphere Networking documentation.
Jumbo Frames
If you plan to use jumbo frames with vSAN to improve CPU performance, verify that jumbo frames are enabled on all
network devices and hosts in the cluster.
By default, the TCP segmentation offload (TSO) and large receive offload (LRO) features are enabled on ESXi. Consider
whether using jumbo frames improves the performance enough to justify the cost of enabling them on all nodes on the
network.
74
VMware vSAN 8.0
In traditional configurations, where vSphere uses a single default gateway, all routed traffic attempts to reach its
destination through this gateway.
NOTE
vSAN 7.0 and later enables you to override the default gateway for the vSAN VMkernel adapter on each host,
and configure a gateway address for the vSAN network.
However, certain vSAN deployments might require static routing. For example, deployments where the witness is on
a different network, or the vSAN stretched cluster deployment, where both the data sites and the witness host are on
different networks.
To configure static routing on your ESXi hosts, use the esxcli command:
esxcli network ip route ipv4 add -g gateway-to-use –n remote-network
remote-network is the remote network that your host must access, and gateway-to-use is the interface to use when
traffic is sent to the remote network.
For information about network design for vSAN stretched clusters, see Administering VMware vSAN.
If a host is not a member of a fault domain, vSAN interprets it as a stand-alone fault domain.
75
VMware vSAN 8.0
76
VMware vSAN 8.0
Preparing Storage
Provide enough disk space for vSAN and for the virtualized workloads that use the vSAN datastore.
77
VMware vSAN 8.0
The storage devices must meet the following requirements so that vSAN can claim them:
• The storage devices are local to the ESXi hosts. vSAN cannot claim remote devices.
• The storage devices do not have any existing partition information.
• On the same host, you cannot have both all-flash and hybrid disk groups.
78
VMware vSAN 8.0
Policy changes • The Failures to tolerate (FTT) influences the physical storage space
that you must supply for virtual machines. The greater the FTT is for
higher availability, the more space you must provide.
When FTT is set to 1, it imposes two replicas of the VMDK file of a
virtual machine. With FTT set to 1, a VMDK file that is 50 GB requires
100-GB space on different hosts. If the FTT is changed to 2, you must
have enough space to support three replicas of the VMDK across the
hosts in the cluster, or 150 GB.
• Some policy changes, such as a new number of disk stripes per object,
require temporary resources. vSAN recreates the objects affected by
the change. For a certain time, the physical storage must accommodate
the old and new objects.
Available space for reprotecting or maintenance mode When you place a host in maintenance mode or you clone a virtual
machine, the datastore might not be able to evacuate the virtual machine
objects, although the vSAN datastore indicates that enough space is
available. This lack of space can occur if the free space is on the host that
is being placed in maintenance mode.
Required mode • Review the vSAN requirements in the VMware Compatibility Guide for the
required mode, passthrough or RAID 0, of the controller.
• If both passthrough and RAID 0 modes are supported, configure passthrough
mode instead of RAID0. RAID 0 introduces complexity for disk replacement.
RAID mode • In the case of RAID 0, create one RAID volume per physical disk device.
• Do not enable a RAID mode other than the mode listed in the VMware
Compatibility Guide.
• Do not enable controller spanning.
Driver and firmware version • Use the latest driver and firmware version for the controller according to
VMware Compatibility Guide.
• If you use the in-box controller driver, verify that the driver is certified for vSAN.
OEM ESXi releases might contain drivers that are not certified and listed in the
VMware Compatibility Guide.
Queue depth Verify that the queue depth of the controller is 256 or higher. Higher queue depth
provides improved performance.
Cache Deactivate the storage controller cache, or set it to 100 percent read if disabling
cache is not possible.
Advanced features Deactivate advanced features, for example, HP SSD Smart Path.
79
VMware vSAN 8.0
Options Description
-d|--disk=str The name of the device that you want to tag as a capacity device. For example, mpx.vmhba1
:C0:T4:L0
-t|--tag=str Specify the tag that you want to add or remove. For example, the capacityFlash tag is
used for marking a flash device for capacity.
80
VMware vSAN 8.0
[
\{
"Name" : "mpx.vmhba1:C0:T4:L0",
"VSANUUID" : "",
"State" : "Eligible for use by VSAN",
"ChecksumSupport": "0",
"Reason" : "None",
"IsSSD" : "1",
"IsCapacityFlash": "0",
"IsPDL" : "0",
\},
For example, to use the same vCenter Server to mark flash devices for capacity as a user root, run the following
command:
rvc root@localhost
81
VMware vSAN 8.0
82
VMware vSAN 8.0
vSAN provides a distributed storage solution, which implies exchanging data across the ESXi hosts that participate in the
cluster. Preparing the network for installing vSAN includes certain configuration aspects.
For information about network design guidelines, see Designing the vSAN Network.
83
VMware vSAN 8.0
• In hybrid clusters, the magnetic disks are used for capacity and flash devices for read and write cache. vSAN allocates
70 percent of all available cache for read cache and 30 percent of available cache for the write buffer. In a hybrid
configuration, the flash devices serve as a read cache and a write buffer.
• In all-flash clusters, one designated flash device is used as a write cache, additional flash devices are used for
capacity. In all-flash clusters, all read requests come directly from the flash pool capacity.
• Only local or direct-attached capacity devices can participate in a vSAN cluster. vSAN cannot consume other external
storage, such as SAN or NAS, attached to cluster.
To learn about the characteristics of a vSAN cluster configured through Quickstart, see Using Quickstart to Configure and
Expand a vSAN Cluster.
For best practices about designing and sizing a vSAN cluster, see Designing and Sizing a vSAN Cluster.
84
VMware vSAN 8.0
Requirements Description
ESXi hosts • Verify that you are using the latest version of ESXi on your hosts.
• Verify that there are at least three ESXi hosts with supported storage configurations
available to be assigned to the vSAN cluster. For best results, configure the vSAN
cluster with four or more hosts.
Memory • Verify that each host has a minimum of 32 GB of memory.
• For larger configurations and better performance, you must have a minimum of 32 GB
of memory in the cluster. See Designing and Sizing vSAN Hosts.
Storage I/O controllers, drivers, firmware • Verify that the storage I/O controllers, drivers, and firmware versions are certified and
listed in the VCG website at http://www.vmware.com/resources/compatibility/search.p
hp.
• Verify that the controller is configured for passthrough or RAID 0 mode.
• Verify that the controller cache and advanced features are deactivated. If you cannot
deactivate the cache, you must set the read cache to 100 percent.
• Verify that you are using controllers with higher queue depths. Using controllers with
queue depths less than 256 can significantly impact the performance of your virtual
machines during maintenance and failure.
Cache and capacity • For vSAN Original Storage Architecture, verify that vSAN hosts contributing storage to
the cluster have at least one cache and one capacity device. vSAN requires exclusive
access to the local cache and capacity devices of the hosts in the vSAN cluster.
They cannot share these devices with other uses, such as Virtual Flash File System
(VFFS), VMFS partitions, or an ESXi boot partition.
• For vSAN Express Storage Architecture, verify that hosts contributing storage have
compatible flash storage devices.
• For best results, create a vSAN cluster with uniformly configured hosts.
Network connectivity • Verify that each host is configured with at least one network adapter.
• For hybrid configurations, verify that vSAN hosts have a minimum dedicated
bandwidth of 1 GbE.
• For all-flash configurations, verify that vSAN hosts have a minimum bandwidth of 10
GbE.
For best practices and considerations about designing the vSAN network, see Designing
the vSAN Network and Networking Requirements for Virtual SAN.
vSAN and vCenter Server compatibility Verify that you are using the latest version of the vCenter Server.
License key • Verify that you have a valid vSAN license key.
• To use the all-flash feature, your license must support that capability.
• To use advanced features, such as vSAN stretched clusters or deduplication and
compression, your license must support those features.
• Verify that the amount of license capacity that you plan on using equals the total
number of CPUs in the hosts participating in the vSAN cluster. Do not provide
license capacity only for hosts providing capacity to the cluster. For information about
licensing for vSAN, see the vCenter Server and Host Management documentation.
For detailed information about vSAN cluster requirements, see Requirements for Enabling vSAN.
For in-depth information about designing and sizing the vSAN cluster, see the VMware vSAN Design and Sizing Guide.
85
VMware vSAN 8.0
Quickstart consolidates the workflow to enable you to quickly configure a new vSAN cluster that uses recommended
default settings for common functions such as networking, storage, and services. Quickstart groups common tasks and
uses configuration wizards that guide you through the process. Once you enter the required information on each wizard,
Quickstart configures the cluster based on your input.
Quickstart uses the vSAN health service to validate the configuration and help you correct configuration issues. Each
Quickstart card displays a configuration checklist. You can click a green message, yellow warning, or red failure to display
details.
Hosts added to a Quickstart cluster are automatically configured to match the cluster settings. The ESXi software and
patch levels of new hosts must match those in the cluster. Hosts cannot have any networking or vSAN configuration
86
VMware vSAN 8.0
when added to a cluster using the Quickstart workflow. For more information about adding hosts, see "Expanding a vSAN
Cluster" in Administering VMware vSAN.
NOTE
If you modify any network settings outside of QuickStart, this hampers your ability to add and configure more
hosts to the cluster using the QuickStart workflow.
Skipping Quickstart
You can use the Skip Quickstart button to exit the Quickstart workflow, and continue configuring the cluster and its hosts
manually. You can add new hosts individually, and manually configure those hosts. Once skipped, you cannot restore the
Quickstart workflow for the cluster.
The Quickstart workflow is designed for new clusters. When you upgrade an existing vSAN cluster to 6.7 Update 1 or
later, the Quickstart workflow appears. Skip the Quickstart workflow and continue to manage the cluster through vCenter
Server.
87
VMware vSAN 8.0
NOTE
If you perform network configuration through Quickstart, then modify those parameters from outside of
Quickstart, you cannot use Quickstart to add or configure additional hosts.
1. Navigate to the cluster in the vSphere Client.
2. Click the Configure tab, and select Configuration > Quickstart.
3. (optional) On the Cluster basics card, click Edit to open the Cluster basics wizard.
a) (Optional) Enter a cluster name.
b) Select basic services, such as DRS, vSphere HA, and vSAN.
Check Enable vSAN ESA to use vSAN Express Storage Architecture. vSAN Express Storage Architecture is
optimized for high-performance flash storage devices that provide greater performance and efficiency.
c) Click OK or Finish.
4. On the Add hosts card, click Add to open the Add hosts wizard.
a) On the Add hosts page, enter information for new hosts, or click Existing hosts and select from hosts listed in the
inventory.
b) On the Host summary page, verify the host settings.
c) On the Ready to complete page, click Finish.
NOTE
If you are running vCenter Server on a host, the host cannot be placed into maintenance mode as you add it
to a cluster using the Quickstart workflow. The same host also can be running a Platform Services Controller.
All other VMs on the host must be powered off.
5. On the Cluster configuration card, click Configure to open the Cluster configuration wizard.
a) (vSAN ESA clusters) On the Cluster Type page, enter the HCI cluster type:
• vSAN HCI provides compute resources and storage resources. The datastore can be shared across data
centers and vCenters.
• vSAN Max provides storage resources, but not compute resources. The datastore can be mounted by remote
vSAN clusters across data centers and vCenters.
b) On the Configure the distributed switches page, enter networking settings, including distributed switches, port
groups, and physical adapters.
• In the Distributed switches section, enter the number of distributed switches to configure from the drop-down
menu. Enter a name for each distributed switch. Click Use Existing to select an existing distributed switch.
If the host has a standard virtual switch with the same name as the selected distributed switch, the standard
switch is migrated to the corresponding distributed switch.
Network resource control is enabled and set to version 3. Distributed switches with network resource control
version 2 cannot be used.
• In the Port Groups section, select a distributed switch to use for vMotion and a distributed switch to use for the
vSAN network.
• In the Physical adapters section, select a distributed switch for each physical network adapter. You must
assign each distributed switch to at least one physical adapter.
If the physical adapters chosen are attached to a standard virtual switch with the same name across hosts, the
standard switch is migrated to the distributed switch. If the physical adapters chosen are unused, there is no
migration from standard switch to distributed switch.
88
VMware vSAN 8.0
Network resource control is enabled and set to version 3. Distributed switches with network resource control
version 2 cannot be used.
c) On the vMotion traffic page, enter IP address information for vMotion traffic.
d) On the Storage traffic page, enter IP address information for storage traffic.
e) On the Advanced options page, enter information for cluster settings, including DRS, HA, vSAN, host options, and
EVC.
f) On the Claim disks page, select storage devices on each host. For clusters with vSAN Original Storage
Architecure, select one cache device and one or more capacity devices. For clusters with vSAN Express STorage
Architecture, select flash devices for the host's storage pool.
NOTE
Only the vSAN Data Persistence platform can consume vSAN Direct storage. The vSAN Data
Persistence platform provides a framework for software technology partners to integrate with VMware
infrastructure. Each partner must develop their own plug-in for VMware customers to receive the benefits
of the vSAN Data Persistence platform. The platform is not operational until the partner solution running
on top is operational. For more information, see vSphere with Tanzu Configuration and Management.
g) (Optional) On the Create fault domains page, define fault domains for hosts that can fail together.
For more information about fault domains, see "Managing Fault Domains in vSAN Clusters" in Administering
VMware vSAN.
h) (Optional) On the Proxy setting page, configure the proxy server if your system uses one.
i) On the Review page, verify the cluster settings, and click Finish.
You can manage the cluster through your vCenter.
You can add hosts to the cluster through Quickstart. For more information. see "Expanding a vSAN Cluster" in
Administering VMware vSAN.
89
VMware vSAN 8.0
90
VMware vSAN 8.0
Create a cluster and add hosts to the cluster before enabling and configuring vSAN. Configure the port properties on each
host to add the vSAN service.
NOTE
You can use Quickstart to quickly create and configure a vSAN cluster. For more information, see "Using
Quickstart to Configure and Expand a vSAN Cluster" in vSAN Planning and Deployment .
1. Navigate to an existing host cluster.
2. Click the Configure tab.
3. Under vSAN, select Services.
b) Select a deployment option (Single site vSAN cluster, Two node vSAN cluster, or vSAN stretched cluster).
c) Click Configure to open the Configure vSAN wizard.
Enabling vSAN creates a vSAN datastore and registers the vSAN storage provider. vSAN storage providers are built-in
software components that communicate the storage capabilities of the datastore to vCenter Server.
Verify that the vSAN datastore has been created. See View vSAN Datastore.
Verify that the vSAN storage provider is registered.
92
VMware vSAN 8.0
93
VMware vSAN 8.0
94
VMware vSAN 8.0
• Configure vSAN Network options. For more information, see "Configuring vSAN Network" in vSAN Planning
and Deployment.
• Configure iSCSI target service. For more information, see "Using the vSAN iSCSI Target Service" in
Administering VMware vSAN.
• Configure Data Services, including deduplication and compression, data-at-rest encryption, and data-in-transit
encryption.
• Configure vSAN Data Protection. Before you can use vSAN Data Protection, you must deploy the vSAN
Snapshot Service. For more information, see "Deploying the Snapshot Service Appliance" in Administering
VMware vSAN.
• Configure capacity reservations and alerts. For more information, see "About Reserved Capacity" in vSAN
Monitoring and Troubleshooting.
• Configure advanced options:
– Object Repair Timer
– Site Read Locality for vSAN stretched clusters
– Thin Swap provisioning
– Large Cluster Support for up to 64 hosts
– Automatic Rebalance
• Configure vSAN historical health service.
c) Modify the settings to match your requirements.
3. Click Apply to confirm your selections.
95
VMware vSAN 8.0
96
VMware vSAN 8.0
1. Navigate to Storage.
2. Select the vSAN datastore.
3. Click the Configure tab.
4. Review the vSAN datastore capacity.
The size of the vSAN datastore depends on the number of capacity devices per ESXi host and the number of ESXi
hosts in the cluster. For example, if a host has seven 2 TB for capacity devices, and the cluster includes eight hosts,
the approximate storage capacity is 7 x 2 TB x 8 = 112 TB. When using the all-flash configuration, flash devices are
used for capacity. For hybrid configuration, magnetic disks are used for capacity.
Some capacity is allocated for metadata.
• On-disk format version 1.0 adds approximately 1 GB per capacity device.
• On-disk format version 2.0 adds capacity overhead, typically no more than 1-2 percent capacity per device.
• On-disk format version 3.0 and later adds capacity overhead, typically no more than 1-2 percent capacity
per device. Deduplication and compression with software checksum enabled require additional overhead of
approximately 6.2 percent capacity per device.
Create a storage policy for virtual machines using the storage capabilities of the vSAN datastore. For information, see the
vSphere Storage documentation.
97
VMware vSAN 8.0
Networking Differences
vSAN uses its own logical network. When vSAN and vSphere HA are enabled for the same cluster, the HA interagent
traffic flows over this storage network rather than the management network. vSphere HA uses the management network
only when vSAN is turned off. vCenter Server chooses the appropriate network when vSphere HA is configured on a host.
NOTE
Make sure vSphere HA is not enabled when you enable vSAN on the cluster. Then you can re-enable vSphere
HA.
When a virtual machine is only partially accessible in all network partitions, you cannot power on the virtual machine
or fully access it in any partition. For example, if you partition a cluster into P1 and P2, the VM namespace object is
accessible to the partition named P1 and not to P2. The VMDK is accessible to the partition named P2 and not to P1. In
such cases, the virtual machine cannot be powered on and it is not fully accessible in any partition .
The following table shows the differences in vSphere HA networking whether or not vSAN is used.
If you change the vSAN network configuration, the vSphere HA agents do not automatically acquire the new network
settings. To change the vSAN network, you must re-enable host monitoring for the vSphere HA cluster:
1. Deactivate Host Monitoring for the vSphere HA cluster.
2. Make the vSAN network changes.
3. Right-click all hosts in the cluster and select Reconfigure HA.
4. Reactivate Host Monitoring for the vSphere HA cluster.
98
VMware vSAN 8.0
than 25 percent of the cluster resources. In the same cluster, with the Failures to tolerate policy, the setting must not
be higher than two hosts. If vSphere HA reserves less capacity, failover activity might be unpredictable. Reserving too
much capacity overly constrains the powering on of virtual machines and intercluster vSphere vMotion migrations. For
information about the Percentage of Cluster Resources Reserved policy, see the vSphere Availability documentation.
99
VMware vSAN 8.0
virtual machines while vSAN is off, make sure you migrate virtual machines from vSAN datastore to another datastore
before turning off the vSAN cluster.
1. Navigate to the vSAN cluster.
2. Click the Configure tab.
3. Under vSAN, select Services.
4. Click Turn Off vSAN.
5. On the Turn Off vSAN dialog, confirm your selection.
Witness Host
Each vSAN stretched cluster consists of two data sites and one witness host. The witness host resides at a third site
and contains the witness components of virtual machine objects. The witness host does not store customer data, only
metadata, such as the size and UUID of vSAN object and components.
The witness host serves as a tiebreaker when a decision must be made regarding availability of datastore components
when the network connection between the two sites is lost. In this case, the witness host typically forms a vSAN cluster
with the preferred site. But if the preferred site becomes isolated from the secondary site and the witness, the witness host
forms a cluster using the secondary site. When the preferred site is online again, data is resynchronized to ensure that
both sites have the latest copies of all data.
If the witness host fails, all corresponding objects become noncompliant but are fully accessible.
100
VMware vSAN 8.0
101
VMware vSAN 8.0
102
VMware vSAN 8.0
103
VMware vSAN 8.0
Two-node vSAN clusters are often used for remote office/branch office environments, typically running a small number of
workloads that require high availability. A two-node vSAN cluster consists of two hosts at the same location, connected to
the same network switch or directly connected. A third host acts as a witness host, which can be located remotely from
the branch office. Usually the witness host resides at the main site, with the vCenter Server.
A single witness host can support up to 64 two-node vSAN clusters. The number of clusters supported by a shared
witness host is based on the host memory.
When you configure a two-node vSAN cluster in Quickstart or with the Configure vSAN wizard, you can select a witness
host. To assign a new witness host for your cluster, right-click the cluster in the vSphere Client and select menu vSAN >
Assign Shared Witness.
104
VMware vSAN 8.0
Network resource control is enabled and set to version 3. Distributed switches with network resource control
version 2 cannot be used.
• In the Port Groups section, select a distributed switch to use for vMotion and a distributed switch to use for the
vSAN network.
• In the Physical adapters section, select a distributed switch for each physical network adapter. You must
assign each distributed switch to at least one physical adapter.
This mapping of physical NICs to the distributed switches is applied to all hosts in the cluster. If you are using
an existing distributed switch, the physical adapter selection can match the mapping of the distributed switch.
c) On the vMotion traffic page, enter IP address information for vMotion traffic.
d) On the Storage traffic page, enter IP address information for storage traffic.
e) On the Advanced options page, enter information for cluster settings, including DRS, HA, vSAN, host options, and
EVC.
In the vSAN options section, select vSAN Stretched cluster or Two node vSAN cluster as the Deployment type.
f) On the Claim disks page, select storage devices to create the vSAN datastore.
For vSAN Original Storage Architecture, select devices for cache and for capacity. vSAN uses those devices to
create disk groups on each host.
For vSAN Express Storage Architecture, select compatible flash devices or enable I want vSAN to manage the
disks. vSAN uses those devices to create storage pools on each host.
g) (Optional) On the Proxy settings, page, configure the proxy server if your system uses one.
h) On the Configure fault domains page, define fault domains for the hosts in the Preferred site and the Secondary
site.
For more information about fault domains, see "Managing Fault Domains in vSAN Clusters" in Administering
VMware vSAN.
i) On the Select witness host page, select a host to use as a witness host. The witness host cannot be part of the
vSAN stretched cluster, and it can have only one VMkernel adapter configured for vSAN data traffic.
Before you configure the witness host, verify that it is empty and does not contain any components. A two-node
vSAN cluster can share a witness with other two-node vSAN clusters.
j) On the Claim disks for witness host page, select disks on the witness host.
k) On the Review page, verify the cluster settings, and click Finish.
You can manage the cluster through vCenter Server.
You can add hosts to the cluster and modify the configuration through Quickstart. You also can modify the configuration
manually with the vSphere Client.
105
VMware vSAN 8.0
4. Click Configure Stretched Cluster to open the vSAN stretched cluster configuration wizard.
5. Select the hosts that you want to assign to the secondary fault domain and click >>.
The hosts that are listed under the Preferred fault domain are in the preferred site.
6. Click Next.
7. Select a witness host that is not a member of the vSAN stretched cluster and click Next.
8. Claim storage devices on the witness host and click Next.
For vSAN Original Storage Architecture, select devices for cache and for capacity.
For vSAN Express Storage Architecture, select compatible flash devices or enable I want vSAN to manage the
disks.
106
VMware vSAN 8.0
9. On the Ready to complete page, review the configuration and click Finish.
107
VMware vSAN 8.0
2. Deploy the appliance to a vSAN host or cluster. For more information, see Deploying OVF Templates in the vSphere
Virtual Machine Administration documentation.
3. Configure the vSAN network on the witness appliance.
4. Configure the management network on the witness appliance.
5. Add the appliance to vCenter Server as a witness ESXi host. Make sure to configure the vSAN VMkernel interface on
the host.
108
VMware vSAN 8.0
vmk1
Name: vmk1
MAC Address: 00:50:56:6a:3a:74
109
VMware vSAN 8.0
Enabled: true
Portset: vSwitch1
Portgroup: vsandata
Netstack Instance: defaultTcpipStack
VDS Name: N/A
VDS UUID: N/A
VDS Port: N/A
VDS Connection: -1
Opaque Network ID: N/A
Opaque Network Type: N/A
External ID: N/A
MTU: 9000
TSO MSS: 65535
Port ID: 50331660
NOTE
Multicast information is included for backward compatibility. vSAN 6.6 and later releases do not require
multicast.
3. Use the esxcli vsan network ip add command to configure the management VMkernel network adapter to support
witness traffic.
esxcli vsan network ip add -i vmkx -T witness
4. Use the esxcli vsan network list command to verify the new network configuration.
For example:
esxcli vsan network list
Interface
VmkNic Name: vmk0
IP Protocol: IP
Interface UUID: 8cf3ec57-c9ea-148b-56e1-a0369f56dcc0
Agent Group Multicast Address: 224.2.3.4
Agent Group IPv6 Multicast Address: ff19::2:3:4
Agent Group Multicast Port: 23451
Master Group Multicast Address: 224.1.2.3
Master Group IPv6 Multicast Address: ff19::1:2:3
Master Group Multicast Port: 12345
Host Unicast Channel Bound Port: 12321
Multicast TTL: 5
Traffic Type: witness
Interface
VmkNic Name: vmk1
IP Protocol: IP
Interface UUID: 6df3ec57-4fb6-5722-da3d-a0369f56dcc0
Agent Group Multicast Address: 224.2.3.4
Agent Group IPv6 Multicast Address: ff19::2:3:4
Agent Group Multicast Port: 23451
Master Group Multicast Address: 224.1.2.3
Master Group IPv6 Multicast Address: ff19::1:2:3
Master Group Multicast Port: 12345
Host Unicast Channel Bound Port: 12321
Multicast TTL: 5
Traffic Type: vsan
110
VMware vSAN 8.0
In the vSphere Client, the management VMkernel network interface is not selected for vSAN traffic. Do not re-enable the
interface in the vSphere Client.
111
VMware vSAN 8.0
Intended Audience
This guide is intended for anyone who is designing, deploying, and managing a vSAN cluster. The information in this
guide is written for experienced network administrators who are familiar with network design and configuration, virtual
machine management, and virtual data center operations. This guide also assumes familiarity with VMware vSphere,
including VMware ESXi, vCenter Server, and the vSphere Client.
Related Documents
In addition to this guide, you can refer to the following guides to know more about vSAN networking:
• vSAN Planning and Deployment Guide, to know more about creating vSAN clusters
• Administering VMware vSAN, to configure a vSAN cluster and learn more about vSAN features
• vSAN Monitoring and Troubleshooting Guide, to monitor and troubleshoot vSAN clusters
112
VMware vSAN 8.0
Terms Definitions
113
VMware vSAN 8.0
Terms Definitions
114
VMware vSAN 8.0
Virtual machines (VMs) on vSAN datastores are made up of a set of objects, and each object can be made up of one or
more components. These components are distributed across multiple hosts for resilience to drive and host failures. vSAN
maintains and updates these components using the vSAN network.
The following diagram provides a high-level overview of the vSAN network:
115
VMware vSAN 8.0
Each ESXi host in a vSAN cluster must have a network adapter for vSAN communication. All the intra-cluster node
communication happens through the vSAN VMkernel port. VMkernel ports provide Layer 2 and Layer 3 services to each
vSAN host and hosted virtual machines.
vSAN Network Traffic
Several different traffic types are available in the vSAN network, such as the storage traffic and the unicast traffic. The
compute and storage of a virtual machine can be on the same host or on different hosts in the cluster. A VM that is not
configured to tolerate a failure might be running on one host, and accessing a VM object or component that resides on a
different host. This implies that all I/O from the VM passes through the network. The storage traffic constitutes most of the
traffic in a vSAN cluster.
The cluster-related communication between all the ESXi hosts creates traffic in the vSAN cluster. This unicast traffic also
contributes to the vSAN network traffic.
Virtual Switch
vSAN supports the following types of virtual switches:
• The Standard Virtual Switch provides connectivity from VMs and VMkernel ports to external networks. This switch is
local to each ESXi host.
• A vSphere Distributed Switch provides central control of the virtual switch administration across multiple ESXi hosts. A
distributed switch also provides networking features such as Network I/O Control (NIOC) that can help you set Quality
of Service (QoS) levels on vSphere or virtual network. vSAN includes vSphere Distributed Switch irrespective of the
vCenter Server version.
Bandwidth
vSAN traffic can share physical network adapters with other system traffic types, such as vSphere vMotion traffic, vSphere
HA traffic, and virtual machine traffic. It also provides more bandwidth for shared network configurations where vSAN,
vSphere management, vSphere vMotion traffic, and so on, are on the same physical network. To guarantee the amount of
bandwidth required for vSAN, use vSphere Network I/O Control in the distributed switch.
In vSphere Network I/O Control, you can configure reservation and shares for the vSAN outgoing traffic:
• Set a reservation so that Network I/O Control guarantees that a minimum bandwidth is available on the physical
adapter for vSAN.
• Set the share value to 100 so that when the physical adapter assigned for vSAN becomes saturated, certain bandwidth
is available to vSAN. For example, the physical adapter might become saturated when another physical adapter in the
team fails and all traffic in the port group is transferred to the other adapters in the team.
For information about using Network I/O Control to configure bandwidth allocation for vSAN traffic, see the vSphere
Networking documentation.
116
VMware vSAN 8.0
Management network The management network is the primary network interface that
uses a VMkernel TCP/IP stack to facilitate the host connectivity
and management. It can also handle the system traffic such as
vMotion, iSCSI, Network File System (NFS), Fiber Channel over
Ethernet (FCoE), and fault tolerance.
Virtual Machine network With virtual networking, you can network virtual machines and
build complex networks within a single ESXi host or across
multiple ESXi hosts.
vMotion network Traffic type that facilitates migration of VM from one host to
another. Migration with vMotion requires correctly configured
network interfaces on source and target hosts. Ensure that the
vMotion network is distinct from the vSAN network.
vSAN network A vSAN cluster requires the VMkernel network for the exchange of
data. Each ESXi host in the vSAN cluster must have a VMkernel
network adapter for the vSAN traffic. For more information, refer to
"Manually Enabling vSAN" in vSAN Planning and Deployment.
Latency Bandwidth
Inter- Between Between
Topology or Support for
Support for Support for Inter-Node Site Link Nodes Nodes
Deployment Architecture NICs Greater
1 GbE NIC 10 GbE NIC Latency Bandwidth and vSAN and vSAN
Mode than 10 GbE
or Latency Witness Witness
Hosts Hosts
Single Site Hybrid Yes Yes Yes Less than 1 NA NA NA
vSAN Cluster Cluster (Minimum) (Recommended) ms RTT.
All-Flash No Yes Yes
Cluster (Recommended)
vSAN Hybrid or All- No Yes Yes Less than RecommendedLess than 2 Mbps
Stretched Flash Cluster (Minimum) 1 ms RTT is 10 GbE 200 ms RTT. per 1000
Cluster within each (Workload Up to 10 components
site. Dependent) hosts per site. (Maximum
and 5 ms Less than of 100 Mbps
RTT or less. 100 ms RTT. with 45 k
11–15 hosts components).
per site.
117
VMware vSAN 8.0
Latency Bandwidth
Inter- Between Between
Topology or Support for
Support for Support for Inter-Node Site Link Nodes Nodes
Deployment Architecture NICs Greater
1 GbE NIC 10 GbE NIC Latency Bandwidth and vSAN and vSAN
Mode than 10 GbE
or Latency Witness Witness
Hosts Hosts
Two-Node Hybrid Yes (Up to 10 Yes Yes Less than RecommendedLess than 2 Mbps
vSAN Cluster Cluster VMs) (Recommended) 1 ms RTT is 10 GbE 500 ms RTT. per 1000
All-Flash No Yes within the and 5 ms components
Cluster (Minimum) same site. RTT or less. (Maximum of
1.5 Mbps).
Latency Bandwidth
Support for Inter-Site Link Between Between
Deployment Support for 1 Support for 10 Inter-Node
NICs Greater Bandwidth or Nodes and Nodes and
Type GbE NIC GbE NIC Latency
than 10 GbE Latency vSAN Witness vSAN Witness
Hosts Hosts
Single Site No Yes Yes Less than 1 ms NA NA NA
vSAN Cluster RTT.
vSAN No Yes Yes Less than 1 Minimum 10 Less than 200 2 Mbps
Stretched ms RTT within GbE (workload ms RTT. Up to per 1000
Cluster each site. dependent) and 10 hosts per components
5 ms RTT. site. (Maximum
Less than 100 of 100 Mbps
ms RTT. 11–15 with 45 k
hosts per site. components).
Two-Node No Yes Yes Less than 1 ms Recommended Less than 500 2 Mbps
vSAN Cluster RTT within the is 25 GbE and ms RTT. per 1000
same site. 5 ms RTT or components
less. (Maximum of
1.5 Mbps).
NOTE
These NIC requirements assume that the packet loss is not more than 0.0001% in the hyper-converged
environments. There can be a drastic impact on the vSAN performance, if any of these requirements are
exceeded.
For more information about the vSAN stretched cluster NIC requirements, see vSAN Stretched Cluster Guide.
118
VMware vSAN 8.0
Site to Site vSAN OSA: minimum of 10 Gbps Less than 5 ms latency RTT.
vSAN ESA: minimum of 10 Gbps
Site to Witness 2 Mbps per 1000 vSAN components • Less than 500 ms latency RTT for 1
host per site.
• Less than 200 ms latency RTT for up to
10 hosts per site.
• Less than 100 ms latency RTT for 11-15
hosts per site.
119
VMware vSAN 8.0
Site to Witness Witness Traffic Separation Layer 3 Use static routes or gateway
override when using an
interface other than the
Management (vmk0) interface.
Site to Witness Witness Traffic Separation Layer 2 for two-host cluster Static routes are not required.
Firewall Considerations
When you enable vSAN on a cluster, all required ports are added to ESXi firewall rules and configured automatically.
There is no need for an administrator to open any firewall ports or enable any firewall services manually.
120
VMware vSAN 8.0
You can view open ports for incoming and outgoing connections. Select the ESXi host, and click Configure > Security
Profile.
121
VMware vSAN 8.0
on-disk format. To ensure that vSAN cluster continues communicating in unicast mode and does not revert to multicast,
upgrade the disk groups on the vSAN 6.6 hosts to on-disk version 5.0.
NOTE
Avoid having a mixed mode cluster where vSAN version 6.5 or earlier are available in the same cluster along
with vSAN version 6.6 or later.
Cluster Information
Enabled: true
Current Local Time: 2020-04-09T18:19:52Z
Local Node UUID: 5e8e3dc3-43ab-5452-795b-a03d6f88f022
Local Node Type: NORMAL
Local Node State: AGENT
Local Node Health State: HEALTHY
Sub-Cluster Master UUID: 5e8e3d3f-3015-9075-49b6-a03d6f88d426
Sub-Cluster Backup UUID: 5e8e3daf-e5e0-ddb6-a523-a03d6f88dd4a
122
VMware vSAN 8.0
NodeUuid IsWitness Supports Unicast IP Address Port Iface Name Cert Thumbprint
SubClusterUuid
------------------------------------ --------- ---------------- ---------- ----- ----------
5e8e3d73-6d1c-0b81-1305-a03d6f888d22 0 true 10.198.95.10 12321
43:80:B7:A1:3F:D1:64:07:8C:58:01:2B:CE:A2:F5:DE:D6:B1:41:AB
5e8e3daf-e5e0-ddb6-a523-a03d6f88dd4a 0 true 10.198.94.240 12321
FE:39:D7:A5:EF:80:D6:41:CD:13:70:BD:88:2D:38:6C:A0:1D:36:69
5e8e3d3f-3015-9075-49b6-a03d6f88d426 0 true 10.198.94.244 12321
72:A3:80:36:F7:5D:8F:CE:B0:26:02:96:00:23:7D:8E:C5:8C:0B:E1
5e8e3d33-5825-ee5c-013c-a03d6f88ea4c 0 true 10.198.95.11 12321
5A:55:74:E8:5F:40:2F:2B:09:B5:42:29:FF:1C:95:41:AB:28:E0:57
The output includes the vSAN node UUID, IPv4 address, IPv6 address, UDP port with which vSAN node communicates,
and whether the node is a data host (0) or a witness host (1). You can use this output to identify the vSAN cluster nodes
that are operating in unicast mode and view the other hosts in the cluster. vCenter Server maintains the output list.
Interface
VmkNic Name: vmk1
IP Protocol: IP
Interface UUID: e290be58-15fe-61e5-1043-246e962c24d0
Agent Group Multicast Address: 224.2.3.4
Agent Group IPv6 Multicast Address: ff19::2:3:4
Agent Group Multicast Port: 23451
Master Group Multicast Address: 224.1.2.3
Master Group IPv6 Multicast Address: ff19::1:2:3
Master Group Multicast Port: 12345
123
VMware vSAN 8.0
Intra-Cluster Traffic
In unicast mode, the primary node addresses all the cluster nodes as it sends the same message to all the vSAN nodes in
a cluster.
For example, if N is the number of vSAN nodes, then the primary node sends the messages N number of times. This
results in a slight increase of vSAN CMMDS traffic. You might not notice this slight increase of traffic during normal,
steady-state operations.
124
VMware vSAN 8.0
With the unicast traffic, there is no change in the witness site traffic requirements.
125
VMware vSAN 8.0
NOTE
vSAN does not have its own TCP/IP stack. Use static routes to route vSAN traffic across L3 networks.
vSphere 6.0 introduced a new TCP/IP stack architecture, which can use multiple TPC/IP stacks to manage different
VMkernel network interfaces. With this architecture, you can configure traffic services such as vMotion, management, and
fault tolerance on isolated TCP/IP stacks, which can use multiple default gateways.
For network traffic isolation and security requirements, deploy the different traffic services onto different network segments
or VLANs. This prevents the different traffic services from traversing through the same default gateway.
When you configure the traffic services on separate TCP/IP stacks, deploy each traffic service type onto its own network
segment. The network segments are accessed through a physical network adapter with VLAN segmentation. Map each
segment to different VMkernel network interfaces with the respective traffic services enabled.
126
VMware vSAN 8.0
vSphere RDMA
vSAN 7.0 Update 2 and later supports Remote Direct Memory Access (RDMA) communication.
RDMA allows direct memory access from the memory of one computer to the memory of another computer without
involving the operating system or CPU. The transfer of memory is offloaded to the RDMA-capable Host Channel Adapters
(HCA).
vSAN supports the RoCE v2 protocol. RoCE v2 requires a network configured for lossless operation that is free of
congestion. If your network has congestion, certain large I/O workloads might experience lower performance than TCP.
Each vSAN host must have a vSAN certified RDMA-capable NIC, as listed in the vSAN section of the VMware
Compatibility Guide. Use only the same model network adapters from the same vendor on each end of the connection.
NOTE
vSphere RDMA is not supported on vSAN stretched clusters, two-node vSAN clusters, vSAN Max, or Datastore
Sharing (HCI-mesh).
All hosts in the cluster must support RDMA. If any host loses RDMA support, the entire vSAN cluster switches to TCP.
vSAN with RDMA supports NIC failover, but does not support LACP or IP-hash-based NIC teaming.
IPv6 Support
vSAN 6.2 and later supports IPv6.
Static Routes
You can use static routes to allow vSAN network interfaces from hosts on one subnet to reach the hosts on another
network.
Most organizations separate the vSAN network from the management network, so the vSAN network does not have a
default gateway. In an L3 deployment, hosts that are on different subnets or different L2 segments cannot reach each
other over the default gateway, which is typically associated with the management network.
Use static routes to allow the vSAN network interfaces from hosts on one subnet to reach the vSAN networks on hosts on
the other network. Static routes instruct a host how to reach a particular network over an interface, rather than using the
default gateway.
The following example shows how to add an IPv4 static route to an ESXi host. Specify the gateway (-g) and the network (-
n) you want to reach through that gateway:
esxcli network ip route ipv4 add –g 172.16.10.253 -n 192.168.10.0/24
When the static routes have been added, vSAN traffic connectivity is available across all networks, assuming the physical
infrastructure allows it. Run the vmkping command to test and confirm communication between the different networks by
127
VMware vSAN 8.0
pinging the IP address or the default gateway of the remote network. You also can check different size packets (-s) and
prevent fragmentation (-d) of the packet.
vmkping –I vmk3 192.168.10.253
Jumbo Frames
vSAN fully supports jumbo frames on the vSAN network.
Jumbo frames are Ethernet frames with more than 1500 bytes of payload. Jumbo frames typically carry up to 9000 bytes
of payload, but variations exist.
Using jumbo frames can reduce CPU utilization and improve throughput.
NOTE
Enable jumbo frames support for vSAN Max deployments to improve performance.
You must decide whether these gains outweigh the overhead of implementing jumbo frames throughout the network. In
data centers where jumbo frames are already enabled in the network infrastructure, you can use them for vSAN. The
operational cost of configuring jumbo frames throughout the network might outweigh the limited CPU and performance
benefits.
Flow Control
You can use flow control to manage the rate of data transfer between two devices.
Flow control is configured when two physically connected devices perform auto-negotiation.
An overwhelmed network node might send a pause frame to halt the transmission of the sender for a specified period. A
frame with a multicast destination address sent to a switch is forwarded out through all other ports of the switch. Pause
frames have a special multicast destination address that distinguishes them from other multicast traffic. A compliant switch
does not forward a pause frame. Frames sent to this range are meant to be acted upon only within the switch. Pause
frames have a limited duration, and expire after a time interval. Two computers that are connected through a switch never
send pause frames to each other, but can send pause frames to a switch.
One reason to use pause frames is to support network interface controllers (NICs) that do not have enough buffering to
handle full-speed reception. This problem is uncommon with advances in bus speeds and memory sizes.
128
VMware vSAN 8.0
Congestion Control
Congestion control helps you control the traffic on the network.
Congestion control applies mainly to packet switching networks. Network congestion within a switch might be caused by
overloaded inter-switch links. If inter-switch links overload the capability on the physical layer, the switch introduces pause
frames to protect itself.
129
VMware vSAN 8.0
130
VMware vSAN 8.0
131
VMware vSAN 8.0
Settings: Failback
This option determines how a physical adapter is returned to active duty after recovering from a failure. A failover event
triggers the network traffic to move from one NIC to another. When a link up state is detected on the originating NIC,
traffic automatically reverts to the original network adapter when Failback is set to Yes. When Failback is set to No, a
manual failback is required.
Setting Failback to No can be useful in some situations. For example, after a physical switch port recovers from a failure,
the port might be active but can take several seconds to begin forwarding traffic. Automatic Failback has been known to
cause problems in certain environments that use the Spanning Tree Protocol. For more information about Spanning Tree
Protocol (STP), see VMware KB 1003804.
132
VMware vSAN 8.0
NOTE
If the host and switch are both in passive mode, the LAG does not initialize, because an active part is required to
trigger the linking. At least one must be Active.
In vSphere 5.5 and later releases, this functionality is called Enhanced LACP. This functionality is only supported on
vSphere Distributed Switch version 5.5 or later.
For more information about LACP support on a vSphere Distributed Switch, see the vSphere 6 Networking
documentation.
NOTE
The number of LAGs you can use depends on the capabilities of the underlying physical environment and the
topology of the virtual network.
For more information about the different load-balancing options, see KB 2051826.
133
VMware vSAN 8.0
134
VMware vSAN 8.0
135
VMware vSAN 8.0
Cons
• Physical switch configuration is less flexible. Physical switch ports must be configured in a static port-channel
configuration.
• Increased chance of misconfiguration. Static port-channels form without any verification on either end (unlike LACP
dynamic port-channel).
• More complex. Introducing full physical redundancy configuration increases complexity when multiple switches are
used. Implementations can become quite vendor specific.
• Limited load balancing. If your environment has only a few IP addresses, the virtual switch might consistently pass
the traffic through one uplink in the team. This can be especially true for small vSAN clusters.
136
VMware vSAN 8.0
Cons
• vSAN does not support multiple VMkernel adapters (vmknics) on the same subnet. For more information, see VMware
KB 2010877.
• Setup is complex and error prone, so troubleshooting is more complex.
• Network availability is not guaranteed with multiple vmknics in some asymmetric failures, such as one NIC failure on
one host and another NIC failure on another host.
• Load-balanced vSAN traffic across physical NICs is not guaranteed.
• Costs increase for vSAN hosts, as you might need multiple VMkernel adapters (vmknics) to protect multiple physical
NICs (vmnics). For example, 2 x 2 vmnics might be required to provide redundancy for two vSAN vmknics.
• Required logical resources are doubled, such as VMkernel ports, IP addresses, and VLANs.
• vSAN does not implement port binding. This means that techniques such as multi-pathing are not available.
• Layer 3 topologies are not suitable for vSAN traffic with multiple vmknics. These topologies might not function as
expected.
• Command-line host configuration might be required to change vSAN multicast addresses.
Dynamic LACP combines, or aggregates, multiple network connections in parallel to increase throughput and provide
redundancy. When NIC teaming is configured with LACP, load balancing of the vSAN network across multiple uplinks
occurs. This load balancing happens at the network layer, and is not done through vSAN.
NOTE
Other terms sometimes used to describe link aggregation include port trunking, link bundling, Ethernet/network/
NIC bonding, EtherChannel.
This section focuses on Link Aggregation Control Protocol (LACP). The IEEE standard is 802.3ad, but some vendors
have proprietary LACP features, such as PAgP (Port Aggregation Protocol). Follow the best practices recommended by
your vendor.
NOTE
The LACP support introduced in vSphere Distributed Switch 5.1 only supports IP-hash load balancing. vSphere
Distributed Switch 5.5 and later fully support LACP.
LACP is an industry standard that uses port-channels. Many hashing algorithms are available. The vSwitch port-group
policy and the port-channel configuration must agree and match.
137
VMware vSAN 8.0
Load Balancing
Since this is a single VMkernel NIC, there is no performance benefit to using Route based on physical load.
Only one physical NIC is in use at any time. The other physical NIC is idle.
138
VMware vSAN 8.0
For each port group, the teaming and failover policy use the default settings.
• Load balancing set to Route based on originating port ID
• Network failure detection set to Link Status Only
• Notify Switches set to the default value of Yes
• Failback set to the default value of Yes
• The uplink configuration has one uplink in the Active position and one uplink in the Unused position.
One network is completely isolated from the other network.
139
VMware vSAN 8.0
Load Balancing
vSAN has no load balancing mechanism to differentiate between multiple vmknics, so the vSAN I/O path chosen is not
deterministic across physical NICs. The vSphere performance charts show that one physical NIC is often more utilized
than the other. A simple I/O test performed in our labs, using 120 VMs with a 70:30 read/write ratio with a 64K block size
on a four-host all flash vSAN cluster, revealed an unbalanced load across NICs.
vSphere performance graphs show an unbalanced load across NICs.
140
VMware vSAN 8.0
You might have to modify disk timeout settings of some guest OSes to ensure that they are not severely impacted. Disk
timeout values might vary, depending on the presence of VMware Tools, and the specific guest OS type and version. For
more information about changing guest OS disk timeout values, go to VMware KB 1009465.
Configure vSphere
Configure the vSphere network with the following settings.
• Configure vDS with the correct MTU.
• Add hosts to vDS.
• Create a LAG with the correct number of uplinks and matching attributes to port channel.
• Assign physical uplinks to the LAG.
• Create a distributed port group for vSAN traffic and assign correct VLAN.
• Configure VMkernel ports for vSAN with correct MTU.
141
VMware vSAN 8.0
Use the following commands to configure an individual port channel on a Dell switch:
• Create a port-channel.
# interface port-channel 1
• Set port-channel to VLAN trunk mode.
# switchport mode trunk
• Allow VLAN access.
# switchport trunk allowed vlan 3262
• Configure the load balancing option.
# hashing-mode 6
• Assign the correct ports to the port-channel and set the mode to Active.
• Verify that the port channel is configured correctly.
# show interfaces port-channel 1
Channel Ports Ch-Type Hash Type Min-links Local Prf
------- ----------------------------- -------- --------- --------- ---------
Po1 Active: Te1/0/36, Te1/0/18 Dynamic 6 1 Disabled
Hash Algorithm Type
1 - Source MAC, VLAN, EtherType, source module and port Id
2 - Destination MAC, VLAN, EtherType, source module and port Id
3 - Source IP and source TCP/UDP port
4 - Destination IP and destination TCP/UDP port
5 - Source/Destination MAC, VLAN, EtherType, source MODID/port
6 - Source/Destination IP and source/destination TCP/UDP port
7 - Enhanced hashing mode
# interface range Te1/0/36, Te1/0/18
# channel-group 1 mode active
Full configuration:
# interface port-channel 1
# switchport mode trunk
# switchport trunk allowed vlan 3262
# hashing-mode 6
# exit
# interface range Te1/0/36,Te1/018
# channel-group 1 mode active
# show interfaces port-channel 1
NOTE
Repeat this procedure on all participating switch ports that are connected to vSAN hosts.
142
VMware vSAN 8.0
143
VMware vSAN 8.0
144
VMware vSAN 8.0
145
VMware vSAN 8.0
In this configuration, use 10Gb networking with two physical uplinks per server. A single VMkernel interface (vmknic) for
vSAN exists on each host.
For more information about host requirements and configuration examples, see the following VMware Knowledge Base
articles:
• Host requirements for link aggregation for ESXi and ESX (1001938)
• Sample configuration of EtherChannel / Link Aggregation Control Protocol (LACP) with ESXi/ESX and Cisco/HP
switches (KB 1004048)
NOTE
vSAN over RDMA does not support this configuration.
146
VMware vSAN 8.0
147
VMware vSAN 8.0
Network Redundancy
In this example, vmnic1 is connected to a port that has been disabled from the switch, to focus on failure and redundancy
behavior. Note that a network uplink redundancy alarm has triggered.
No vSAN health alarms were triggered. Cluster and VM components are not affected and Guest Storage I/O is not
interrupted by this failure.
148
VMware vSAN 8.0
149
VMware vSAN 8.0
Edit the other traffic types to match the share values shown in the table.
If the 10 GbE adapter becomes saturated, Network I/O Control allocates 5 Gbps to vSAN on the physical adapter,
3.5 Gbps to virtual machine traffic, and 1.5 Gbps to vMotion. Use these values as a starting point to configure NIOC
configuration on your vSAN network. Ensure that vSAN has the highest priority of any protocol.
For more details about the various parameters for bandwidth allocation, see vSphere Networking documentation.
With each of the vSphere editions for vSAN, VMware provides a vSphere Distributed Switch as part of the edition.
Network I/O Control can be configured with any vSAN edition.
Standard Deployments
vSAN supports several single-site deployment types.
150
VMware vSAN 8.0
The Layer 2 network topology offers the simplest implementation and management of vSAN. VMware recommends
the use and configuration of IGMP Snooping to avoid sending unnecessary multicast traffic on the network. In this first
example, we are looking at a single site, and perhaps even a single rack of servers using vSAN 6.5 or earlier. This
version uses multicast, so enable IGMP Snooping. Since everything is on the same L2, you need not configure routing for
multicast traffic.
Layer 2 implementations are simplified even further with vSAN 6.6 and later, which introduces unicast support. IGMP
Snooping is not required.
151
VMware vSAN 8.0
152
VMware vSAN 8.0
Layer 2 Everywhere
You can configure a vSAN stretched cluster in a Layer 2 network, but this configuration is not recommended.
Consider a design where the vSAN stretched cluster is configured in one large Layer 2 design. Data Site 1 and Site 2 are
where the virtual machines are deployed. Site 3 contains the witness host.
NOTE
For best results, do not use a stretched Layer 2 network across all sites.
To demonstrate Layer 2 everywhere as simply as possible, we use switches (and not routers) in the topologies.
Layer 2 networks cannot have any loops (multiple paths), so features such as Spanning Tree Protocol (STP) are needed
to block one of the connections between Site 1 and Site 2. Now consider a situation where the link between Site 2 and
Site 3 is broken (the link between Site 1 and Site 2). Network traffic can be switched from Site 1 to Site 2 through the
witness host at Site 3. As VMware supports a much lower bandwidth and higher latency for the witness host, you see a
significant decrease in performance if data network traffic passes through a lower specification witness site.
If switching traffic between data sites through the witness site does not impact latency of applications, and bandwidth is
acceptable, a stretched L2 configuration between sites is possible. In most cases, such a configuration is not feasible, and
adds complexity to the networking requirements.
153
VMware vSAN 8.0
With vSAN 6.5 or earlier, which uses multicast traffic, you must configure IGMP snooping on the switches. This is not
necessary with vSAN 6.6 and later. PIM is not necessary because there is no routing of multicast traffic.
154
VMware vSAN 8.0
With vSAN 6.6, which uses unicast, there is no requirement to consider IGMP snooping or PIM.
Layer 3 Everywhere
In this vSAN stretched cluster configuration, the data traffic is routed between the data sites and the witness host.
To implement Layer 3 everywhere as simply as possible, use routers or routing switches in the topologies.
For example, consider an environment with vSAN 6.5 or earlier, which uses multicast traffic. In this case, configure IGMP
snooping on the data site switches to manage the amount of multicast traffic on the network. This is unnecessary at the
witness host since witness traffic is unicast. The multicast traffic is routed between the data sites, so configure PIM to
allow multicast routing.
With vSAN 6.6 and later, neither IGMP snooping nor PIM are needed because all the routed traffic is unicast.
155
VMware vSAN 8.0
156
VMware vSAN 8.0
In this example, if rack 1 fails, rack 2 and the witness host provide VM availability. This configuration is a pre- vSAN 6.6
environment, and needs multicast configured on the network. The witness host must be on the vSAN network. Witness
traffic is unicast. In vSAN 6.6 and later, all traffic is unicast.
This topology is also supported over L3. Place the vSAN VMkernel ports on different subnets or VLANs, and use a
separate subnet or VLAN for each rack.
157
VMware vSAN 8.0
This topology supports deployments with two racks to achieve rack awareness (fault domains) with a vSAN stretched
cluster. This solution uses a witness host that is external to the cluster.
158
VMware vSAN 8.0
To enable this functionality, the witness traffic is separated completely from the vSAN data traffic. The vSAN data traffic
can flow between the two nodes on the direct connect, while the witness traffic can be routed to the witness site over the
management network.
The witness appliance can be located remotely from the branch office. For example, the witness might be running back in
the main data center, alongside the management infrastructure (vCenter Server, vROps, Log Insight, and so on). Another
supported place where the witness can reside remotely from the branch office is in vCloud Air.
In this configuration, there is no switch at the remote site. As a result, there is no need configure support for multicast
traffic on the vSAN back-to-back network. You do not need to consider multicast on the management network because the
witness traffic is unicast.
vSAN 6.6 and later uses all unicast, so there are no multicast considerations. Multiple remote office/branch office two-
node deployments are also supported, so long as each has their own unique witness.
159
VMware vSAN 8.0
160
VMware vSAN 8.0
Option 2: Virtual ESXi Witness Appliance Connected over L3 with Static Routes
Since the witness host is a virtual machine that gets deployed on a physical ESXi host, which is not part of the vSAN
cluster, that physical ESXi host must have a minimum of one VM network pre-configured. This VM network must reach
both the management network and the vSAN network shared by the ESXi hosts on the data sites.
NOTE
The witness host does not need to be a dedicated host. It can be used for many other VM workloads, while
simultaneously hosting the witness.
An alternative option is to have two preconfigured VM networks on the underlying physical ESXi host, one for the
management network and one for the vSAN network. When the virtual ESXi witness is deployed on this physical ESXi
host, the network needs to be attached and configured accordingly.
Once you have deployed the virtual ESXi witness host, configure the static route. Assume that the data sites are
connected over a stretched L2 network. Use this also for the data sites’ management network, vSAN network, vMotion
network, and virtual machine network. vSAN traffic is not routed from the hosts in the data sites (site 1 and site 2) to the
host in the witness site (site 3) over the default gateway. To configure the vSAN stretched cluster successfully, all hosts in
the cluster require static routes, so that the vSAN traffic from site 1 and site 2 can reach the witness host in site 3. Use the
esxcli network ip route command to add a static route on each ESXi host.
161
VMware vSAN 8.0
hosts. If latency is greater than this value, consider a vSAN stretched cluster which tolerates latency of 5ms. With vSAN
6.5 or earlier, additional considerations for multicast must be addressed.
For best results, maintain a uniform configuration across all sites in such a topology. To maintain availability of VMs,
configure fault domains, where the hosts in each room, building, or site are placed in the same fault domain. Avoid
asymmetric partitioning of the cluster, where host A cannot communicate to host B, but host B can communicate to host A.
162
VMware vSAN 8.0
• vMotion: MTU checks (ping with large packet size). This check complements the basic vMotion ping connectivity
check. Maximum Transmission Unit size is increased to improve network performance. Incorrectly configured MTUs
might not appear as a network configuration issue, but can cause performance issues.
• vSAN cluster partition. This health check examines the cluster to see how many partitions exist. It displays an error if
there is more than a single partition in the vSAN cluster.
• Multicast assessment based on other checks. This health check aggregates data from all network health checks. If
this check fails, it indicates that multicast is likely the root cause of a network partition.
This provides useful information, such as which VMkernel interface is being used for vSAN traffic. In this case, it is vmk1.
However, also shown are the multicast addresses. This information might be displayed even when the cluster us running
in unicast mode. There is the group multicast address and port. Port 23451 is used for the heartbeat, sent every second
by the primary, and is visible on every other host in the cluster. Port 12345 is used for the CMMDS updates between the
primary and backup.
163
VMware vSAN 8.0
The Maximum Transmission Unit size is shown as 9000, so this VMkernel port is configured for jumbo frames, which
require an MTU of about 9,000. VMware does not make any recommendation around the use of jumbo frames. However,
jumbo frames are supported for use with vSAN.
vmkping
The vmkping command verifies whether all the other ESXi hosts on the network are responding to your ping requests.
~ # vmkping -I vmk2 172.32.0.3 -s 1472 -d
PING 172.32.0.3 (172.32.0.3): 56 data bytes
64 bytes from 172.32.0.3: icmp_seq=0 ttl=64 time=0.186 ms
64 bytes from 172.32.0.3: icmp_seq=1 ttl=64 time=2.690 ms
64 bytes from 172.32.0.3: icmp_seq=2 ttl=64 time=0.139 ms
While it does not verify multicast functionality, it can help identify a rogue ESXi host that has network issues. You can also
examine the response times to see if there is any abnormal latency on the vSAN network.
164
VMware vSAN 8.0
If jumbo frames are configured, this command does not report any issues if the jumbo frame MTU size is incorrect. By
default, this command uses an MTU size of 1500. If there is a need to verify if jumbo frames are successfully working end-
to-end, use vmkping with a larger packet size (-s) option as follows:
~ # vmkping -I vmk2 172.32.0.3 -s 8972 -d
PING 172.32.0.3 (172.32.0.3): 8972 data bytes
9008 bytes from 172.32.0.3: icmp_seq=0 ttl=64 time=0.554 ms
9008 bytes from 172.32.0.3: icmp_seq=1 ttl=64 time=0.638 ms
9008 bytes from 172.32.0.3: icmp_seq=2 ttl=64 time=0.533 ms
Consider adding -d to the vmkping command to test if packets can be sent without fragmentation.
Received Bytes: 64
Host: 172.40.0.10
ICMP Seq: 1
TTL: 64
Round-trip Time: 1834 us
165
VMware vSAN 8.0
Dup: false
Detail:
Received Bytes: 64
Host: 172.40.0.10
ICMP Seq: 2
TTL: 64
Round-trip Time: 1824 us
Dup: false
Detail:
Summary:
Host Addr: 172.40.0.10
Transmitted: 3
Recieved: 3
Duplicated: 0
Packet Lost: 0
Round-trip Min: 1824 us
Round-trip Avg: 1840 us
Round-trip Max: 1864 us
[root@esxi-dell-m:~]
vsan.lldpnetmap
This RVC command displays uplink port information.
If there are non-Cisco switches with Link Layer Discovery Protocol (LLDP) enabled in the environment, there is an
RVC command to display uplink <-> switch <-> switch port information. For more information on RVC, refer to the RVC
Command Guide.
This helps you determine which hosts are attached to which switches when the vSAN cluster is spanning multiple
switches. It can help isolate a problem to a particular switch when only a subset of the hosts in the cluster is impacted.
> vsan.lldpnetmap 02013-08-15 19:34:18 -0700: This operation will take 30-60 seconds ...+---------------
+---------------------------+| Host | LLDP info |+---------------
+---------------------------+| 10.143.188.54 | w2r13-vsan-x650-2: vmnic7 || | w2r13-vsan-x650-1:
vmnic5 |+---------------+---------------------------+
This is only available with switches that support LLDP. To configure it, log in to the switch and run the following:
switch# config t
Switch(Config)# feature lldp
NOTE
LLDP operates in both send and receive mode, by default. Check the settings of your vDS properties if the
physical switch information is not being discovered. By default, vDS is created with discovery protocol set to
CDP, Cisco Discovery Protocol. To resolve this, set the discovery protocol to LLDP, and set operation to both on
the vDS.
166
VMware vSAN 8.0
One of the simplest ways to verify if multicast is working correctly in your vSAN environment is by using the tcpdump-uw
command. This command is available from the command line of the ESXi hosts.
This tcpdump-uw command shows if the primary is correctly sending multicast packets (port and IP info) and if all other
hosts in the cluster are receiving them.
On the primary, this command shows the packets being sent out to the multicast address. On all other hosts, the same
packets are visible (from the primary to the multicast address). If they are not visible, multicast is not working correctly.
Run the tcpdump-uw command shown here on any host in the cluster, and the heartbeats from the primary are visible. In
this case, the primary is at IP address 172.32.0.2. The -v for verbosity is optional.
[root@esxi-hp-02:~] tcpdump-uw -i vmk2 multicast -v
tcpdump-uw: listening on vmk2, link-type EN10MB (Ethernet), capture size 96 bytes
11:04:21.800575 IP truncated-ip - 146 bytes missing! (tos 0x0, ttl 5, id 34917, offset 0, flags [none], proto
UDP (17), length 228)
172.32.0.4.44824 > 224.1.2.3.12345: UDP, length 200
11:04:22.252369 IP truncated-ip - 234 bytes missing! (tos 0x0, ttl 5, id 15011, offset 0, flags [none], proto
UDP (17), length 316)
172.32.0.2.38170 > 224.2.3.4.23451: UDP, length 288
11:04:22.262099 IP truncated-ip - 146 bytes missing! (tos 0x0, ttl 5, id 3359, offset 0, flags [none], proto
UDP (17), length 228)
172.32.0.3.41220 > 224.2.3.4.23451: UDP, length 200
11:04:22.324496 IP truncated-ip - 146 bytes missing! (tos 0x0, ttl 5, id 20914, offset 0, flags [none], proto
UDP (17), length 228)
172.32.0.5.60460 > 224.1.2.3.12345: UDP, length 200
11:04:22.800782 IP truncated-ip - 146 bytes missing! (tos 0x0, ttl 5, id 35010, offset 0, flags [none], proto
UDP (17), length 228)
172.32.0.4.44824 > 224.1.2.3.12345: UDP, length 200
11:04:23.252390 IP truncated-ip - 234 bytes missing! (tos 0x0, ttl 5, id 15083, offset 0, flags [none], proto
UDP (17), length 316)
172.32.0.2.38170 > 224.2.3.4.23451: UDP, length 288
11:04:23.262141 IP truncated-ip - 146 bytes missing! (tos 0x0, ttl 5, id 3442, offset 0, flags [none], proto
UDP (17), length 228)
172.32.0.3.41220 > 224.2.3.4.23451: UDP, length 200
While this output might seem a little confusing, suffice to say that the output shown here indicates that the four hosts in the
cluster are getting a heartbeat from the primary. This tcpdump-uw command must be run on every host to verify that they
are all receiving the heartbeat. This verifies that the primary is sending the heartbeats, and every other host in the cluster
is receiving them, which indicates that multicast is working.
If some of the vSAN hosts are not able to pick up the one-second heartbeats from the primary, the network administrator
needs to check the multicast configuration of their switches.
To avoid the annoying truncated-ip – 146 bytes missing! message, use the –s0 option to the same command
to stop trunacating of packets:
[root@esxi-hp-02:~] tcpdump-uw -i vmk2 multicast -v -s0
tcpdump-uw: listening on vmk2, link-type EN10MB (Ethernet), capture size 65535 bytes
11:18:29.823622 IP (tos 0x0, ttl 5, id 56621, offset 0, flags [none], proto UDP (17), length 228)
172.32.0.4.44824 > 224.1.2.3.12345: UDP, length 200
11:18:30.251078 IP (tos 0x0, ttl 5, id 52095, offset 0, flags [none], proto UDP (17), length 228)
172.32.0.3.41220 > 224.2.3.4.23451: UDP, length 200
11:18:30.267177 IP (tos 0x0, ttl 5, id 8228, offset 0, flags [none], proto UDP (17), length 316)
172.32.0.2.38170 > 224.2.3.4.23451: UDP, length 288
11:18:30.336480 IP (tos 0x0, ttl 5, id 28606, offset 0, flags [none], proto UDP (17), length 228)
172.32.0.5.60460 > 224.1.2.3.12345: UDP, length 200
167
VMware vSAN 8.0
11:18:30.823669 IP (tos 0x0, ttl 5, id 56679, offset 0, flags [none], proto UDP (17), length 228)
172.32.0.4.44824 > 224.1.2.3.12345: UDP, length 200
The tcpdump command is related to IGMP (Internet Group Management Protocol) membership. Hosts (and network
devices) use IGMP to establish multicast group membership.
Each ESXi host in the vSAN cluster sends out regular IGMP membership reports (Join).
The tcpdump command shows IGMP member reports from a host:
[root@esxi-dell-m:~] tcpdump-uw -i vmk1 igmp
tcpdump-uw: verbose output suppressed, use -v or -vv for full protocol decode
listening on vmk1, link-type EN10MB (Ethernet), capture size 262144 bytes
15:49:23.134458 IP 172.40.0.9 > igmp.mcast.net: igmp v3 report, 1 group record(s)
15:50:22.994461 IP 172.40.0.9 > igmp.mcast.net: igmp v3 report, 1 group record(s)
The output shows IGMP v3 reports are taking place, indicating that the ESXi host is regularly updating its membership. If
a network administrator has any doubts whether or not vSAN ESXi hosts are doing IGMP correctly, running this command
on each ESXi host in the cluster and showing this trace can be used to verify.
If you have multicast communications, use IGMP v3.
In fact, the following command can be used to look at multicast and IGMP traffic at the same time:
[root@esxi-hp-02:~] tcpdump-uw -i vmk2 multicast or igmp -v -s0
A common issue is that the vSAN cluster is configured across multiple physical switches, and while multicast has been
enabled on one switch, it has not been enabled across switches. In this case, the cluster forms with two ESXi hosts in one
partition, and another ESXi host (connected to the other switch) is unable to join this cluster. Instead it forms its own vSAN
cluster in another partition. The vsan.lldpnetmap command seen earlier can help you determine network configuration,
and which hosts are attached to which switch.
While a vSAN cluster forms, there are indicators that show multicast might be an issue.
Assume that the checklist for subnet, VLAN, MTU has been followed, and each host in the cluster can vmkping every
other host in the cluster.
If there is a multicast issue when the cluster is created, a common symptom is that each ESXi host forms its own vSAN
cluster, with itself as the primary. If each host has a unique network partition ID, this symptom suggests that there is no
multicast between any of the hosts.
However, if there is a situation where a subset of the ESXi hosts form a cluster, and another subset form another cluster,
and each have unique partitions with their own primary, backup and perhaps even agent hosts, multicast is enabled in
the switch, but not across switches. vSAN shows hosts on the first physical switch forming their own cluster partition, and
hosts on the second physical switch forming their own cluster partition, each with its own primary. If you can verify which
switches the hosts in the cluster connect to, and hosts in a cluster are connected to the same switch, then this probably is
the issue.
168
VMware vSAN 8.0
169
VMware vSAN 8.0
From a network perspective, it is the RDT associations (Assocs) and sockets count that are important. There are 45,000
associations per host in vSAN 6.0 and later. An RDT association is used to track peer-to-peer network state within vSAN.
vSAN is sized so that it never runs out of RDT associations. vSAN also limits how many TCP sockets it is allowed to use,
and vSAN is sized so that it never runs out of its allocation of TCP sockets. There is a limit of 10,000 sockets per host.
A vSAN client represents object's access in the vSAN cluster. The client typically represents a virtual machine running on
a host. The client and the object might not be on the same host. There is no hard defined limit, but this metric is shown to
help understand how clients balance across hosts.
There is only one vSAN owner for a given vSAN object, typically co-located with the vSAN client accessing this object.
vSAN owners coordinate all access to the vSAN object and implement functionality, such as mirroring and striping. There
is no hard defined limit, but this metric is once again shown to help understand how owners balance across hosts.
170
VMware vSAN 8.0
NOTE
You can deactivate IGMP snooping if vSAN is on a non-routed or trunked VLAN that you can extend to the
vSAN ports of all the hosts in the cluster.
Server Message Block (SMB) TCP port 445 File Servers External network to file servers
Quotas for a user of a local TCP port 875 File Servers External network to file servers
filesystem (RQUOTA)
Network File System (NFS) TCP and UDP port 2049 File Servers External network to file servers.
NFSv3 can use both TCP and
UDP ports but NFSv4.1 uses
only TCP.
171
VMware vSAN 8.0
NFS Mount TCP and UDP port 20048 File Servers External network to file servers
Network Status Monitor (NSM) TCP and UDP port 27689 File Servers External network to file servers.
server daemon Both inward and outward
communication must be
permitted.
Network Lock Manager (NLM) TCP and UDP port 32803 File Servers External network to file servers.
Allows the connection initiated
from File Server to client.
Inbound and outbound
connections must be allowed on
firewall. The default port is UDP.
Sun remote procedure call TCP and UDP port 111 File Servers External network to file servers
(sunrpc)
LDAP TCP port 389 Active Directory (AD) servers (if File servers to AD servers
AD domain is configured)
LDAP to Global Catalog TCP port 3268 AD servers (if AD domain is File servers to AD servers
configured)
Kerberos TCP port 88 AD servers (if AD domain is File servers to AD servers
configured)
Kerberos password change TCP port 464 AD servers (if AD domain is File servers to AD servers
configured)
Domain Name Server (DNS) TCP and UDP port 53 DNS servers File servers to DNS servers
vSAN Distributed File System TCP port 1564 ESXi hosts Inside vSAN network
(VDFS) Server
Remote Procedure Call TCP port 135 AD servers (if AD domain is File servers to AD servers
configured)
NetBIOS Session Service TCP port 139 AD servers (if AD domain is File servers to AD servers
configured)
DNS UDP port 53 AD servers (if AD domain is File servers to AD servers
configured)
LDAP, DC Locator, and Net Log UDP port 389 AD servers (if AD domain is File servers to AD servers
on configured)
Randomly allocated high TCP TCP 49152 - 65535 AD servers (if AD domain is File servers to AD servers
ports configured)
172
VMware vSAN 8.0
vSAN identifies each iSCSI target by a unique iSCSI qualified Name (IQN). The iSCSI target is presented to a remote
iSCSI initiator using the IQN, so that the initiator can access the LUN of the target. vSAN iSCSI target service allows
creating iSCSI initiator groups. The iSCSI initiator group restricts access to only those initiators that are members of the
group.
173
VMware vSAN 8.0
174
VMware vSAN 8.0
3. Click OK.
Create distributed port groups for vMotion, virtual machine networking, and vSAN networking.
1. Right-click the vSphere Distributed Switch and select menu Distributed Port Group > New Distributed Port Group.
2. For this example, create a port group for the vMotion network.
Create all the distributed port groups on the distributed vSwitch. Then migrate the uplinks, VMkernel networking, and
virtual machine networking to the distributed vSwitch and associated distributed port groups.
WARNING
Migrate the uplinks and networks in step-by-step fashion to proceed smoothly and with caution.
175
VMware vSAN 8.0
Migrate vMotion
To migrate the vMotion network, use the same steps used for the management network.
Before you begin, ensure that the distributed port group for the vMotion network has the same attributes as the port group
on the standard vSwitch. Then migrate the uplink used for vMotion (vmnic1), with the VMkernel adapter (vmk1).
176
VMware vSAN 8.0
Migrate VM Network
The final task needed to migrate the network from a standard vSwitch to a distributed vSwitch is to migrate the VM
network.
Manage host networking.
1. Right-click the vDS and choose menu Add and Manage Hosts.
2. Select all the hosts in the cluster, to migrate virtual machine networking for all hosts to the distributed vSwitch.
Do not move any uplinks. However, if the VM networking on your hosts used a different uplink, then migrate the uplink
from the standard vSwitch.
3. Select the VMs to migrate from a virtual machine network on the standard vSwitch to the virtual machine distributed
port group on the distributed vSwitch. Click Assign port group, and select the distributed port group.
4. Review the changes and click Finish. In this example, you are moving to VMs. Any templates using the original
standard vSwitch virtual machine network must be converted to virtual machines, and edited. The new distributed
port group for virtual machines must be selected as the network. This step cannot be achieved through the migration
wizard.
Since the standard vSwitch no longer has any uplinks or port groups, it can be safely removed.
This completes the migration from a vSphere Standard Switch to a vSphere Distributed Switch.
177
VMware vSAN 8.0
• If your vSAN version is earlier than v6.6 and multiple vSAN clusters are on the same network, configure multicast to
use unique multicast addresses.
• If your vSAN version is earlier than v6.6 and spans multiple switches, verify that multicast is configured across
switches.
• If your vSAN version is earlier than v6.6 and is routed, verify that PIM is configured to allow multicast routing.
• Ensure that the physical switch can meet vSAN requirements (multicast, flow control, feature interoperability).
• Verify that the network does not have performance issues, such as excessive dropped packets or pause frames.
• Verify that network limits are within acceptable margins.
• Test vSAN network performance with iperf, and verify that it meets expectations.
178
VMware vSAN 8.0
Intended Audience
This information is for experienced virtualization administrators who are familiar with virtualization technology, day-to-day
data center operations, and vSAN concepts.
For more information about vSAN and how to create a vSAN cluster, see the vSAN Planning and Deployment Guide.
For more information about monitoring a vSAN cluster and fixing problems, see the vSAN Monitoring and Troubleshooting
Guide.
Updated Information
This document is updated with each release of the product or when necessary.
This table provides the update history of Administering VMware vSAN.
Revision Description
What Is vSAN
VMware vSAN is a distributed layer of software that runs natively as a part of the ESXi hypervisor.
vSAN aggregates local or direct-attached capacity devices of a host cluster and creates a single storage pool shared
across all hosts in the vSAN cluster. While supporting VMware features that require shared storage, such as HA, vMotion,
and DRS, vSAN eliminates the need for external shared storage and simplifies storage configuration and virtual machine
provisioning activities.
vSAN Concepts
VMware vSAN uses a software-defined approach that creates shared storage for virtual machines.
It virtualizes the local physical storage resources of ESXi hosts and turns them into pools of storage that can be
divided and assigned to virtual machines and applications according to their quality-of-service requirements. vSAN is
implemented directly in the ESXi hypervisor.
179
VMware vSAN 8.0
You can configure vSAN to work as either a hybrid or all-flash cluster. In hybrid clusters, flash devices are used for the
cache layer and magnetic disks are used for the storage capacity layer. In all-flash clusters, flash devices are used for
both cache and capacity.
You can activate vSAN on existing host clusters, or when you create a new cluster. vSAN aggregates all local capacity
devices into a single datastore shared by all hosts in the vSAN cluster. You can expand the datastore by adding capacity
devices or hosts with capacity devices to the cluster. vSAN works best when all ESXi hosts in the cluster share similar or
identical configurations across all cluster members, including similar or identical storage configurations. This consistent
configuration balances virtual machine storage components across all devices and hosts in the cluster. Hosts without any
local devices also can participate and run their virtual machines on the vSAN datastore.
In vSAN Original Storage Architecture (OSA), each host that contributes storage devices to the vSAN datastore must
provide at least one device for flash cache and at least one device for capacity. The devices on the contributing host
form one or more disk groups. Each disk group contains one flash cache device, and one or multiple capacity devices for
persistent storage. Each host can be configured to use multiple disk groups.
In vSAN Express Storage Architecture (ESA), all storage devices claimed by vSAN contribute to capacity and
performance. Each host's storage devices claimed by vSAN form a storage pool. The storage pool represents the amount
of caching and capacity provided by the host to the vSAN datastore.
For best practices, capacity considerations, and general recommendations about designing and sizing a vSAN cluster,
see the VMware vSAN Design and Sizing Guide.
Characteristics of vSAN
The following characteristics apply to vSAN, its clusters, and datastores.
vSAN includes numerous features to add resiliency and efficiency to your data computing and storage environment.
Shared storage support vSAN supports VMware features that require shared storage, such as HA,
vMotion, and DRS. For example, if a host becomes overloaded, DRS can
migrate virtual machines to other hosts in the cluster.
On-disk format vSAN on-disk virtual file format provides highly scalable snapshot and
clone management support per vSAN cluster. For information about
the number of virtual machine snapshots and clones supported per
vSAN cluster, refer to the vSphere Configuration Maximumshttps://
configmax.esp.vmware.com/home.
All-flash and hybrid configurations vSAN can be configured for all-flash or hybrid cluster.
Fault domains vSAN supports configuring fault domains to protect hosts from rack or
chassis failures when the vSAN cluster spans across multiple racks or
blade server chassis in a data center.
File service vSAN file service enables you to create file shares in the vSAN datastore
that client workstations or VMs can access.
iSCSI target service vSAN iSCSI target service enables hosts and physical workloads that
reside outside the vSAN cluster to access the vSAN datastore.
vSAN Stretched cluster and Two node vSAN cluster vSAN supports stretched clusters that span across two geographic
locations.
180
VMware vSAN 8.0
Support for Windows Server Failover Clusters (WSFC) vSAN 6.7 Update 3 and later releases support SCSI-3 Persistent
Reservations (SCSI3-PR) on a virtual disk level required by Windows
Server Failover Cluster (WSFC) to arbitrate an access to a shared disk
between nodes. Support of SCSI-3 PRs enables configuration of WSFC
with a disk resource shared between VMs natively on vSAN datastores.
Currently the following configurations are supported:
• Up to 6 application nodes per cluster.
• Up to 64 shared virtual disks per node.
NOTE
Microsoft SQL Server 2012 or later running on Microsoft
Windows Server 2012 or later has been qualified on vSAN.
vSAN health service vSAN health service includes preconfigured health check tests to monitor,
troubleshoot, diagnose the cause of cluster component problems, and
identify any potential risk.
vSAN performance service vSAN performance service includes statistical charts used to monitor IOPS,
throughput, latency, and congestion. You can monitor performance of a
vSAN cluster, host, disk group, disk, and VMs.
Integration with vSphere storage features vSAN integrates with vSphere data management features traditionally used
with VMFS and NFS storage. These features include snapshots, linked
clones, and vSphere Replication.
Virtual Machine Storage Policies vSAN works with VM storage policies to support a VM-centric approach to
storage management.
If you do not assign a storage policy to the virtual machine during
deployment, the vSAN Default Storage Policy is automatically assigned to
the VM.
®
Rapid provisioning vSAN enables rapid provisioning of storage in the vCenter Server during
virtual machine creation and deployment operations.
Deduplication and compression vSAN performs block-level deduplication and compression to save
storage space. When you enable deduplication and compression on a
vSAN all-flash cluster, redundant data within each disk group is reduced.
Deduplication and compression is a cluster-wide setting, but the functions
are applied on a disk group basis. Compression-only vSAN is applied on a
per-disk basis.
Data at rest encryption vSAN provides data at rest encryption. Data is encrypted after all other
processing, such as deduplication, is performed. Data at rest encryption
protects data on storage devices, in case a device is removed from the
cluster.
Data in transit encryption vSAN can encrypt data in transit across hosts in the cluster. When you
enable data-in-transit encryption, vSAN encrypts all data and metadata
traffic between hosts.
SDK support The VMware vSAN SDK is an extension of the VMware vSphere
Management SDK. It includes documentation, libraries and code examples
that help developers automate installation, configuration, monitoring, and
troubleshooting of vSAN.
181
VMware vSAN 8.0
Before you get started with vSAN, review the key vSAN terms and definitions.
Consumed Capacity
Consumed capacity is the amount of physical capacity consumed by one or more virtual machines at any point. Many
factors determine consumed capacity, including the consumed size of your .vmdk files, protection replicas, and so on.
When calculating for cache sizing, do not consider the capacity used for protection replicas.
Object-Based Storage
vSAN stores and manages data in the form of flexible data containers called objects. An object is a logical volume that
has its data and metadata distributed across the cluster. For example, every .vmdk is an object, as is every snapshot.
When you provision a virtual machine on a vSAN datastore, vSAN creates a set of objects comprised of multiple
components for each virtual disk. It also creates the VM home namespace, which is a container object that stores all
metadata files of your virtual machine. Based on the assigned virtual machine storage policy, vSAN provisions and
manages each object individually, which might also involve creating a RAID configuration for every object.
NOTE
If vSAN Express Storage Architecture is enabled, every snapshot is not a new object. A base .vmdk and its
snapshots are contained in one vSAN object. Additionally, in vSAN ESA, digest is backed by vSAN objects.
When vSAN creates an object for a virtual disk and determines how to distribute the object in the cluster, it considers the
following factors:
• vSAN verifies that the virtual disk requirements are applied according to the specified virtual machine storage policy
settings.
• vSAN verifies that the correct cluster resources are used at the time of provisioning. For example, based on the
protection policy, vSAN determines how many replicas to create. The performance policy determines the amount of
flash read cache allocated for each replica and how many stripes to create for each replica and where to place them in
the cluster.
• vSAN continually monitors and reports the policy compliance status of the virtual disk. If you find any noncompliant
policy status, you must troubleshoot and resolve the underlying problem.
NOTE
When required, you can edit VM storage policy settings. Changing the storage policy settings does not affect
virtual machine access. vSAN actively throttles the storage and network resources used for reconfiguration
to minimize the impact of object reconfiguration to normal workloads. When you change VM storage policy
182
VMware vSAN 8.0
settings, vSAN might initiate an object recreation process and subsequent resynchronization. See vSAN
Monitoring and Troubleshooting.
• vSAN verifies that the required protection components, such as mirrors and witnesses, are placed on separate hosts
or fault domains. For example, to rebuild components during a failure, vSAN looks for ESXi hosts that satisfy the
placement rules where protection components of virtual machine objects must be placed on two different hosts, or
across fault domains.
vSAN Datastore
After you enable vSAN on a cluster, a single vSAN datastore is created. It appears as another type of datastore in the list
of datastores that might be available, including Virtual Volume, VMFS, and NFS. A single vSAN datastore can provide
different service levels for each virtual machine or each virtual disk. In vCenter Server®, storage characteristics of the
vSAN datastore appear as a set of capabilities. You can reference these capabilities when defining a storage policy for
virtual machines. When you later deploy virtual machines, vSAN uses this policy to place virtual machines in the optimal
manner based on the requirements of each virtual machine. For general information about using storage policies, see the
vSphere Storage documentation.
A vSAN datastore has specific characteristics to consider.
• vSAN provides a single vSAN datastore accessible to all hosts in the cluster, whether or not they contribute storage to
the cluster. Each host can also mount any other datastores, including Virtual Volumes, VMFS, or NFS.
• You can use Storage vMotion to move virtual machines between vSAN datastores, NFS datastores, and VMFS
datastores.
• Only magnetic disks and flash devices used for capacity can contribute to the datastore capacity. The devices used for
flash cache are not counted as part of the datastore.
183
VMware vSAN 8.0
is compliant. From the Physical Disk Placement tab on the Virtual Disks page, you can verify the virtual machine object
compliance status. For information about troubleshooting a vSAN cluster, see vSAN Monitoring and Troubleshooting.
Witness
A witness is a component that contains only metadata and does not contain any actual application data. It serves as
a tiebreaker when a decision must be made regarding the availability of the surviving datastore components, after a
potential failure. A witness consumes approximately 2 MB of space for metadata on the vSAN datastore when using on-
disk format 1.0, and 4 MB for on-disk format version 2.0 and later.
vSAN maintains a quorum by using an asymmetrical voting system where each component might have more than one
vote to decide the availability of objects. Greater than 50 percent of the votes that make up a VM’s storage object must
be accessible at all times for the object to be considered available. When 50 percent or fewer votes are accessible to all
hosts, the object is no longer accessible to the vSAN datastore. Inaccessible objects can impact the availability of the
associated VM.
vSphere PowerCLI
VMware vSphere PowerCLI adds command-line scripting support for vSAN, to help you automate configuration and
management tasks. vSphere PowerCLI provides a Windows PowerShell interface to the vSphere API. PowerCLI includes
cmdlets for administering vSAN components. For information about using vSphere PowerCLI, see vSphere PowerCLI
Documentation.
184
VMware vSAN 8.0
185
VMware vSAN 8.0
Depending on your requirement, you can deploy vSAN in the following ways.
vSAN ReadyNode
The vSAN ReadyNode is a preconfigured solution of the vSAN software provided by VMware partners, such as Cisco,
Dell, HPE, Fujitsu, IBM, and Supermicro. This solution includes validated server configuration in a tested, certified
hardware form factor for vSAN deployment that is recommended by the server OEM and VMware. For information about
the vSAN ReadyNode solution for a specific partner, visit the VMware Partner website.
186
VMware vSAN 8.0
187
VMware vSAN 8.0
vSphere HA
You can enable vSphere HA and vSAN on the same cluster. As with traditional datastores, vSphere HA provides the same
level of protection for virtual machines on vSAN datastores. This level of protection imposes specific restrictions when
vSphere HA and vSAN interact. For specific considerations about integrating vSphere HA and vSAN, see Using vSAN
and vSphere HA.
Limitations of vSAN
This topic discusses the limitations of vSAN.
When working with vSAN, consider the following limitations:
• vSAN does not support hosts participating in multiple vSAN clusters. However, a vSAN host can access other external
storage resources that are shared across clusters.
• vSAN does not support vSphere DPM and Storage I/O Control.
• vSAN does not support SE Sparse disks.
• vSAN does not support RDM, VMFS, diagnostic partition, and other device access features.
188
VMware vSAN 8.0
189
VMware vSAN 8.0
Create a cluster and add hosts to the cluster before enabling and configuring vSAN. Configure the port properties on each
host to add the vSAN service.
NOTE
You can use Quickstart to quickly create and configure a vSAN cluster. For more information, see "Using
Quickstart to Configure and Expand a vSAN Cluster" in vSAN Planning and Deployment .
1. Navigate to an existing host cluster.
2. Click the Configure tab.
3. Under vSAN, select Services.
b) Select a deployment option (Single site vSAN cluster, Two node vSAN cluster, or vSAN stretched cluster).
c) Click Configure to open the Configure vSAN wizard.
Enabling vSAN creates a vSAN datastore and registers the vSAN storage provider. vSAN storage providers are built-in
software components that communicate the storage capabilities of the datastore to vCenter Server.
Verify that the vSAN datastore has been created. See View vSAN Datastore.
Verify that the vSAN storage provider is registered.
191
VMware vSAN 8.0
192
VMware vSAN 8.0
193
VMware vSAN 8.0
• Configure vSAN Network options. For more information, see "Configuring vSAN Network" in vSAN Planning
and Deployment.
• Configure iSCSI target service. For more information, see "Using the vSAN iSCSI Target Service" in
Administering VMware vSAN.
• Configure Data Services, including deduplication and compression, data-at-rest encryption, and data-in-transit
encryption.
• Configure vSAN Data Protection. Before you can use vSAN Data Protection, you must deploy the vSAN
Snapshot Service. For more information, see "Deploying the Snapshot Service Appliance" in Administering
VMware vSAN.
• Configure capacity reservations and alerts. For more information, see "About Reserved Capacity" in vSAN
Monitoring and Troubleshooting.
• Configure advanced options:
– Object Repair Timer
– Site Read Locality for vSAN stretched clusters
– Thin Swap provisioning
– Large Cluster Support for up to 64 hosts
– Automatic Rebalance
• Configure vSAN historical health service.
c) Modify the settings to match your requirements.
3. Click Apply to confirm your selections.
194
VMware vSAN 8.0
1. Navigate to Storage.
2. Select the vSAN datastore.
3. Click the Configure tab.
4. Review the vSAN datastore capacity.
The size of the vSAN datastore depends on the number of capacity devices per ESXi host and the number of ESXi
hosts in the cluster. For example, if a host has seven 2 TB for capacity devices, and the cluster includes eight hosts,
the approximate storage capacity is 7 x 2 TB x 8 = 112 TB. When using the all-flash configuration, flash devices are
used for capacity. For hybrid configuration, magnetic disks are used for capacity.
Some capacity is allocated for metadata.
• On-disk format version 1.0 adds approximately 1 GB per capacity device.
• On-disk format version 2.0 adds capacity overhead, typically no more than 1-2 percent capacity per device.
• On-disk format version 3.0 and later adds capacity overhead, typically no more than 1-2 percent capacity
per device. Deduplication and compression with software checksum enabled require additional overhead of
approximately 6.2 percent capacity per device.
Create a storage policy for virtual machines using the storage capabilities of the vSAN datastore. For information, see the
vSphere Storage documentation.
195
VMware vSAN 8.0
format, then, before uploading, convert it to stream-optimized format using the vmware-vdiskmanager command‐line
utility. For more information, see Virtual Disk Manager User’s Guide.
• When you upload a vmdk file to a vSAN datastore, the vmdk file inherits the default policy of that datastore. The
vmdk does not inherit the policy of the VM from which it was downloaded. vSAN creates the objects by applying the
vsanDatastore default policy, which is RAID -1. You can change the default policy of the datastore. See Change the
Default Storage Policy for vSAN Datastores .
• You must upload a vmdk file to VM home folder.
1. Navigate to vSAN Datastore.
2. Click the Files tab.
Option Description
Upload Files 1. Select the target folder and click Upload Files. You see a
message informing that you can upload vmdk files only in
VMware stream-optimized format. If you try uploading a
vmdk file in a different format, you see an internal server
error message.
2. Click Upload.
3. Locate the item to upload on the local computer and click
Open.
Upload Folders 1. Select the target folder and click Upload Folder. You see
a message informing that you can upload vmdk files only in
VMware stream-optimized format.
2. Click Upload.
3. Locate the item to upload on the local computer and click
Open.
196
VMware vSAN 8.0
vSAN uses storage providers to supply information about underlying storage to the vCenter Server. This information helps
you to make appropriate decisions about virtual machine placement, and to monitor your storage environment.
197
VMware vSAN 8.0
Capability Description
Failures to tolerate (FTT) Defines the number of host and device failures that a virtual machine object can
tolerate. For n failures tolerated, each piece of data written is stored in n+1 places,
including parity copies if using RAID-5 or RAID-6.
If fault domains are configured, 2n+1 fault domains with hosts contributing capacity
are required. A host which does not belong to a fault domain is considered its own
single-host fault domain.
You can select a data replication method that optimizes for performance or
capacity. RAID-1 (Mirroring) uses more disk space to place the components of
objects but provides better performance for accessing the objects. RAID-5/6
(Erasure Coding) uses less disk space, but performance is reduced. You can select
one of the following options:
• No data redundancy: Specify this option if you do not want vSAN to protect
a single mirror copy of virtual machine objects. This means that your data
is unprotected, and you might lose data when the vSAN cluster encounters
a device failure. The host might experience unusual delays when entering
maintenance mode. The delays occur because vSAN must evacuate the object
from the host for the maintenance operation to complete successfully.
• No data redundancy with host affinity: Specify this option only if you want
to run vSAN Shared Nothing Architecture (SNA) workloads on the vSAN Data
Persistence Platform.
• 1 failure - RAID-1 (Mirroring): Specify this option if your VM object can tolerate
one host or device failure. To protect a 100 GB VM object by using RAID-1
(Mirroring) with an FTT of 1, you consume 200 GB.
• 1 failure - RAID-5 (Erasure Coding): Specify this option if your VM object
can tolerate one host or device failure. For vSAN OSA, to protect a 100 GB
VM object by using RAID-5 (Erasure Coding) with an FTT of 1, you consume
133.33 GB.
NOTE
If you use vSAN Express Storage Architecture, vSAN creates an
optimized RAID-5 format based on the cluster size. If the number
of hosts in the cluster is less than 6, vSAN creates a RAID-5 (2+1)
format. If the number of hosts is equal to or greater than 6, vSAN
creates a RAID-6 (4+1) format. When the cluster size eventually
expands or shrinks, vSAN automatically readjusts the format after 24
hours from the configuration change.
• 2 failures - RAID-1 (Mirroring): Specify this option if your VM object can
tolerate up to two device failures. Since you need to have an FTT of 2 using
RAID-1 (Mirroring), there is a capacity overhead. To protect a 100 GB VM
object by using RAID-1(Mirroring) with an FTT of 2, you consume 300 GB.
• 2 failures - RAID-6 (Erasure Coding): Specify this option if your VM objects
can tolerate up to two device failures. To protect a 100 GB VM object by using
RAID-6 (Erasure Coding) with an FTT of 2, you consume 150 GB. For more
information, refer to Using RAID 5 or RAID 6 Erasure Coding in vSAN Cluster.
• 3 failures - RAID-1 (Mirroring): Specify this option if your VM objects can
tolerate up to three device failures. To protect a 100 GB VM object by using
RAID-1 (Mirroring) with an FTT of 3, you consume 400 GB.
NOTE
If you create a storage policy and you do not specify a value for FTT,
vSAN creates a single mirror copy of the VM objects. It can tolerate a
single failure. However, if multiple component failures occur, your data
might be at risk.
Site disaster tolerance This rule defines whether to use a standard, stretched, or 2-node cluster. If you use
a vSAN stretched cluster, you can define whether data is mirrored at both sites or
only at one site. For a vSAN stretched cluster, you can choose to keep data on the
Preferred or Secondary site for host affinity. 198
• None - standard cluster is the default value. This means that there is no site
disaster tolerance.
• Host mirroring - 2 node cluster defines the number of additional failures that
an object can tolerate after the number of failures defined by FTT is reached.
VMware vSAN 8.0
Capability Description
Encryption services Defines the encryption options for the VMs that you deploy to your datastore.
Choone one of the following options:
• Data-At-Rest encryption: Specify this option if you want to apply encryption to
the data that is stored in your datastore.
• No encryption: Specify this option if you do not want to apply any form of
encryption to your data.
• No preference: Specify this option if you do not want to explicitly apply any
encryption rules. By selecting this option, vSAN applies both rules to your VMs.
Space efficiency Defines the space efficiency options for the VMs that you deploy to your datastore.
Choose one of the following options:
• Deduplication and compression: Specify this option if you want to apply both
deduplication and compression to your data.
• Compression only: Specify this option if you want to apply only compression
to your data.
NOTE
For vSAN Original Storage Architecture, compression is a cluster-
level setting. For vSAN Express Storage Architecture, compresson
only is performed at the object level. This means that you can use
compression for one VM but not for another VM in the same cluster.
• No space efficiency: Specify this option if you do not want to apply
compression to your objects.
• No preference: Specify this option if you do not want to explicitly apply
any space efficiency rules. By selecting this option, vSAN applies all space
efficiency rules to your VMs.
Storage tier Specify the storage tier for all VMs with the defined storage policy. Choose one of
the following options:
• All flash: Specify this option if you want to make your VMs compatible with all-
flash environment.
• Hybrid: Specify this option if you want to make your VMs compatible with only
hybrid environment.
• No preference: Specify this option if you do not want to explicitly apply any
storage tier rules. By selecting this option, vSAN makes the VMs compatible
with both hybrid and all flash environments.
199
VMware vSAN 8.0
Capability Description
Number of disk stripes per object The minimum number of capacity devices across which each replica of a virtual
machine object is striped. A value higher than 1 might result in better performance,
but also results in higher use of system resources.
Default value is 1. Maximum value is 12.
Do not change the default striping value.
In a hybrid environment, the disk stripes are spread across magnetic disks. For an
all-flash configuration, the striping is across flash devices that make up the capacity
layer. Make sure that your vSAN environment has sufficient capacity devices
present to accommodate the request.
IOPS limit for object Defines the IOPS limit for an object, such as a VMDK. IOPS is calculated as the
number of I/O operations, using a weighted size. If the system uses the default
base size of 32 KB, a 64-KB I/O represents two I/O operations.
When calculating IOPS, read and write are considered equivalent, but cache hit
ratio and sequentiality are not considered. If a disk’s IOPS exceeds the limit, I/O
operations are throttled. If the IOPS limit for object is set to 0, IOPS limits are not
enforced.
vSAN allows the object to double the rate of the IOPS limit during the first second
of operation or after a period of inactivity.
Object space reservation Percentage of the logical size of the virtual machine disk (vmdk) object that must
be reserved, or thick provisioned when deploying virtual machines. The following
options are available:
• Thin provisioning (default)
• 25% reservation
• 50% reservation
• 75% reservation
• Thick provisioning
Flash read cache reservation (%) Flash capacity reserved as read cache for the virtual machine object. Specified as
a percentage of the logical size of the virtual machine disk (vmdk) object. Reserved
flash capacity cannot be used by other objects. Unreserved flash is shared fairly
among all objects. Use this option only to address specific performance issues.
You do not have to set a reservation to get cache. Setting read cache reservations
might cause a problem when you move the virtual machine object because the
cache reservation settings are always included with the object.
The Flash Read Cache Reservation storage policy attribute is supported only for
hybrid storage configurations. Do not use this attribute when defining a VM storage
policy for an all-flash cluster or for a vSAN ESA cluster.
Default value is 0%. Maximum value is 100%.
NOTE
By default, vSAN dynamically allocates read cache to storage objects
based on demand. This feature represents the most flexible and the
most optimal use of resources. As a result, typically, you do not need to
change the default 0 value for this parameter.
Note: To increase the value when solving a performance problem,
exercise caution. Over-provisioned cache reservations across several
virtual machines can cause flash device space to be wasted on over-
reservations. These cache reservations are not available to service the
workloads that need the required space at a given time. This space
wasting and unavailability might lead to performance degradation.
200
VMware vSAN 8.0
Capability Description
Disable Object Checksum If the option is set to No, the object calculates checksum information to ensure
the integrity of its data. If this option is set to Yes, the object does not calculate
checksum information.
vSAN uses end-to-end checksum to ensure the integrity of data by confirming that
each copy of a file is exactly the same as the source file. The system checks the
validity of the data during read/write operations, and if an error is detected, vSAN
repairs the data or reports the error.
If a checksum mismatch is detected, vSAN automatically repairs the data by
overwriting the incorrect data with the correct data. Checksum calculation and
error-correction are performed as background operations.
The default setting for all objects in the cluster is No, which means that checksum
is enabled.
NOTE
For vSAN Express Storage Architecture, object checksum is always on
and cannot be deactivated.
Force provisioning If the option is set to Yes, the object is provisioned even if the Failures to tolerate,
Number of disk stripes per object, and Flash read cache reservation policies
specified in the storage policy cannot be satisfied by the datastore. Use this
parameter in bootstrapping scenarios and during an outage when standard
provisioning is no longer possible.
The default No is acceptable for most production environments. vSAN fails to
provision a virtual machine when the policy requirements are not met, but it
successfully creates the user-defined storage policy.
When working with virtual machine storage policies, you must understand how the storage capabilities affect the
consumption of storage capacity in the vSAN cluster. For more information about designing and sizing considerations of
storage policies, refer to "Designing and Sizing a vSAN Cluster" in vSAN Planning and Deployment.
201
VMware vSAN 8.0
The vSAN storage providers report a set of underlying storage capabilities to vCenter Server. They also communicate with
the vSAN layer to report the storage requirements of the virtual machines. For more information about storage providers,
see the vSphere Storage documentation.
vSAN 6.7 and later releases register only one vSAN Storage Provider for all the vSAN clusters managed by the vCenter
Server using the following URL:
https://<VC fqdn>:<VC https port>/vsan/vasa/version.xml
Verify that the storage providers are registered.
1. Navigate to .
2. Click the Configure tab, and click Storage Providers.
Specification Setting
Failures to tolerate 1
Number of disk stripes per object 1
Flash read cache reservation, or flash capacity used for the read 0
cache
Object space reservation 0
NOTE
Setting the Object space reservation to zero means
that the virtual disk is thin provisioned, by default.
Force provisioning No
If you use a vSAN Express Storage Architecture cluster, depending on your cluster size, you can use one of the ESA
policies listed here.
Specification Setting
Failures to tolerate 1
Number of disk stripes per object 1
Flash read cache reservation, or flash capacity used for the read 0
cache
202
VMware vSAN 8.0
Specification Setting
NOTE
RAID-5 in vSAN ESA supports three host clusters. If you enable auto-policy management, the cluster must have
four hosts to use RAID-5.
Specification Setting
Failures to tolerate 2
Number of disk stripes per object 1
Flash read cache reservation, or flash capacity used for the read 0
cache
Object space reservation Thin provisioning
Force provisioning No
NOTE
To use RAID-6, you must have at least six hosts in the cluster.
You can review the configuration settings for the default virtual machine storage policy when you navigate to the VM
Storage Policies > name of the default storage policy > Rule-Set 1: VSAN.
For best results, consider creating and using your own VM storage policies, even if the requirements of the policy are
same as those defined in the default storage policy. In some cases, when you scale up a cluster, you must modify the
default storage policy to maintain compliance with the requirements of the Service Level Agreement for VMware Cloud on
AWS.
When you assign a user-defined storage policy to a datastore, vSAN applies the settings for the user-defined policy on the
specified datastore. Only one storage policy can be the default policy for the vSAN datastore.
203
VMware vSAN 8.0
• You cannot edit the name and description of the default policy, or the vSAN storage provider specification. All other
parameters including the policy rules are editable.
• You cannot delete the default storage policy.
• A default storage policy is assigned when the policy that you assign during virtual machine provisioning does not
include rules specific to vSAN.
204
VMware vSAN 8.0
• Verify that the vSAN storage provider is available. Refer to View vSAN Storage Providers.
• Required privileges: Profile-driven storage.Profile-driven storage view and Profile-driven storage.Profile-driven
storage update
NOTE
Clusters with vSAN Express Storage Architecure can use Auto Policy management. For more information, refer
to What are vSAN Default Storage Policies.
In this policy, you reference storage capabilities supported by the vSAN datastore.
1. Navigate to Policies and Profiles, then click VM Storage Policies.
2. Click Create.
3. On the Name and description page, select a vCenter Server.
4. Type a name and a description for the storage policy and click Next.
5. On the Policy structure page, select Enable rules for "vSAN" storage, and click Next.
6. On the vSAN page, define the policy rule set, and click Next.
a) On the Availability tab, define the Site disaster tolerance and Failures to tolerate.
Availability options define the rules for failures to tolerate, Data locality, and Failure tolerance method.
• Site disaster tolerance defines the type of site failure tolerance used for virtual machine objects.
• Failures to tolerate defines the number of host and device failures that a virtual machine object can tolerate,
and the data replication method.
For example, if you choose Dual site mirroring and 2 failures - RAID-6 (Erasure Coding), vSAN configures the
following policy rules:
• Failures to tolerate: 1
• Secondary level of failures to tolerate: 2
• Data locality: None
205
VMware vSAN 8.0
206
VMware vSAN 8.0
207
VMware vSAN 8.0
NOTE
If you are running vCenter Server on a host, the host cannot be placed into maintenance mode as you add it to
a cluster using the Quickstart workflow. The same host also can be running a Platform Services Controller. All
other VMs on the host must be powered off.
1. Navigate to the cluster in the vSphere Client.
2. Click the Configure tab, and select Configuration > Quickstart.
3. On the Add hosts card, click Launch to open the Add hosts wizard.
a) On the Add hosts page, enter information for new hosts, or click Existing hosts and select from hosts listed in the
inventory.
b) On the Host summary page, verify the host settings.
c) On the Ready to complete page, click Finish.
4. On the Cluster configuration card, click Launch to open the Cluster configuration wizard.
a) On the Configure the distributed switches page, enter networking settings for the new hosts.
b) (optional) On the Claim disks page, select disks on each new host.
c) (optional) On the Create fault domains page, move the new hosts into their corresponding fault domains.
For more information about fault domains, see Managing Fault Domains in vSAN Clusters.
d) On the Ready to complete page, verify the cluster settings, and click Finish.
3. Click Next.
4. View the summary information and click Next.
5. Review the settings and click Finish.
The host is added to the cluster.
Verify that the vSAN Disk Balance health check is green.
For more information about vSAN cluster configuration and fixing problems, see "vSAN Cluster Configuration Issues" in
vSAN Monitoring and Troubleshooting.
208
VMware vSAN 8.0
209
VMware vSAN 8.0
To view specific details about which parameters differ between the host that failed compliance and the host profile,
click the Monitor tab and select the Compliance view. Expand the object hierarchy and select the non-compliant
host. The parameters that differ are displayed in the Compliance window, below the hierarchy.
If compliance fails, use the Remediate action to apply the host profile settings to the host. This action changes all
host profile-managed parameters to the values that are contained in the host profile attached to the host.
c) To view specific details about which parameters differ between the host that failed compliance and the host profile,
click the Monitor tab and select the Compliance view.
d) Expand the object hierarchy and select the failing host.
The parameters that differ are displayed in the Compliance window, below the hierarchy.
5. Remediate the host to fix compliance errors.
a) Select the Monitor tab and click Compliance.
b) Right-click the host or hosts to remediate and select All vCenter Actions > Host Profiles > Remediate.
You can update or change the user input parameters for the host profiles policies by customizing the host.
c) Click Next.
d) Review the tasks that are necessary to remediate the host profile and click Finish.
The host is part of the vSAN cluster and its resources are accessible to the vSAN cluster. The host can also access all
existing vSAN storage I/O policies in the vSAN cluster.
210
VMware vSAN 8.0
You can provision VMs running on your local cluster to use storage space on a remote datastore. When you provision a
new virtual machine, you can select a remote datastore that is mounted to the client cluster. You can ssign any compatible
storage policy configured for the remote datastore.
Mounting a remote datastore is a cluster-wide configuration. When you mount a remote datastore to a vSAN cluster, it is
available to all hosts in the cluster.
When you create a vSphere cluster for mounting remote datastore, select any one of the following vSAN cluster types:
• vSAN HCI cluster provides compute resources and storage resources. It can share its datastore across data centers
and vCenter instances and mount datastores from other vSAN HCI clusters.
• vSAN Compute Cluster is a vSphere cluster that provides compute resources only. It can mount datastores served by
vSAN max clusters and vSAN HCI clusters.
• vSAN Max (vSAN ESA only) provides storage resources, but not compute resources. Its datastore can be mounted by
remote vSAN Max or vSAN HCI clusters across data centers and vCenter instances.
vSAN datastore sharing has the following design considerations:
• vSAN Original Storage Architecture clusters running 8.0 Update 1 or later can share datastores across clusters in the
same data center, or across clusters managed by remote vCenters, as long as they are on the same network. vSAN
Express Storage Architecture clusters running 8.0 Update 2 or later have this feature.
• A vSAN HCI or vSAN Max cluster can serve its local datastore to up to 10 client clusters.
• A client cluster can mount up to 5 remote datastores from one or more vSAN server clusters.
• A single datastore can be mounted to up to 128 vSAN hosts, including hosts in the local vSAN server cluster.
• All objects that make up a VM must reside on the same datastore.
• For vSphere HA to work with vSAN datastore sharing, configure the following failure response for Datastore with APD:
Power off and restart VMs.
• Client hosts that are not part of a cluster are not supported. You can configure a single host compute-only cluster, but
vSphere HA does not work unless you add a second host to the cluster.
• Data-in-transit encryption is not supported.
The following configurations are not supported with vSAN datastore sharing:
• Remote provisioning of iSCSI volumes, or CNS persistent volumes. You can provision iSCSI volumes on the local
vSAN datastore, but not on any remote vSAN datastore. For remote provisioning of CNS persistent volumes, see
vSphere Functionality Supported by vSphere Container Storage Plug-in and Using vSphere Container Storage Plug-in
for HCI Mesh Deployment in the vSphere Storage guide.
• Air-gapped networks or clusters using multiple vSAN VMkernel ports
211
VMware vSAN 8.0
212
VMware vSAN 8.0
Use the Remote Datastores view to monitor and manage remote datastores mounted on the local vSAN cluster. Each
client vSAN cluster can mount remote datastores from server vSAN clusters. Each compatible vSAN cluster also can act
as a server, and allow other vSAN clusters to mount its local datastore.
213
VMware vSAN 8.0
Monitor views for capacity, performance, health, and placement of virtual objects show the status of remote objects and
datastores.
214
VMware vSAN 8.0
Use the vCenter's Remote Datastores page to manage remote datastore sources (Configure > vSAN > Remote
Datastores). Click the tabs to access information about shared datastores across vCenters, add vCenters as datastores
sources, and mount datastores to local clusters.
Datastore Sources View and manage datastore sources residing in remote vCenters. You can add or remove remote datastore
sources for the local vCenter.
Clusters View and manage clusters residing in the local vCenter. You can mount or unmount datastores from remote
vCenters to the selected cluster.
Datastores View all datastores available under this vCenter.
215
VMware vSAN 8.0
This view lists information about each datastore mounted to the local cluster.
• Server cluster that hosts the datastore
• vCenter of the server cluster (if applicable)
• Capacity usage of the datastore
• Free capacity available
• Number of VMs using the datastore (number of VMs using the compute resources of the local cluster, but the storage
resources of the server cluster)
• Client clusters that have mounted the datastore
You can mount or unmount remote datastores from this page.
216
VMware vSAN 8.0
217
VMware vSAN 8.0
You can run pro-active tests on remote datastores to verify VM creation and network performance. The VM creation test
creates a VM on the remote datastore. The Network performance test checks the network performance between all hosts
in the client cluster and all hosts the server clusters.
218
VMware vSAN 8.0
The remote vCenter is added as a datastore source. vSAN clusters on this vCenter can mount remote datastores that
reside on the remote vCenter.
219
VMware vSAN 8.0
The Confirm Maintenance Mode dialog box provides information to guide your maintenance activities. You can view the
impact of each data evacuation option.
• Whether or not sufficient capacity is available to perform the operation.
• How much data will be moved.
• How many objects will become non-compliant.
• How many objects will become inaccessible.
220
VMware vSAN 8.0
221
VMware vSAN 8.0
Any vSAN iSCSI targets served by this host are transferred to other hosts in the cluster, and thus the iSCSI initiator are
redirected to the new target owner.
1. Right-click the host and select Maintenance Mode > Enter Maintenance Mode.
2. Select a data evacuation mode and click OK.
Option Description
Ensure accessibility This is the default option. When you power off or remove the
host from the cluster, vSAN migrates just enough data to ensure
every object is accessible after the host goes into maintenance
mode. Select this option if you want to take the host out of the
cluster temporarily, for example, to install upgrades, and plan to
have the host back in the cluster. This option is not appropriate if
you want to remove the host from the cluster permanently.
Typically, only partial data evacuation is required. However,
the virtual machine might no longer be fully compliant to a VM
storage policy during evacuation. That means, it might not have
access to all its replicas. If a failure occurs while the host is in
maintenance mode and the Failures to tolerate is set to 1, you
might experience data loss in the cluster.
NOTE
This is the only evacuation mode available if you are
working with a three-host cluster or a vSAN cluster
configured with three fault domains.
Full data migration vSAN evacuates all data to other hosts in the cluster and
maintains the current object compliance state. Select this option
if you plan to migrate the host permanently. When evacuating
data from the last host in the cluster, make sure that you migrate
the virtual machines to another datastore and then place the
host in maintenance mode.
This evacuation mode results in the largest amount of data
transfer and consumes the most time and resources. All
the components on the local storage of the selected host
are migrated elsewhere in the cluster. When the host enters
maintenance mode, all virtual machines have access to their
storage components and are still compliant with their assigned
storage policies.
NOTE
If there are objects in reduced availability state, this
mode maintains this compliance state and does not
guarantee that the objects will become compliant.
If a virtual machine object that has data on the host
is not accessible and is not fully evacuated, the host
cannot enter maintenance mode.
222
VMware vSAN 8.0
Option Description
No data migration vSAN does not evacuate any data from this host. If you power
off or remove the host from the cluster, some virtual machines
might become inaccessible.
A cluster with three fault domains has the same restrictions that a three-host cluster has, such as the inability to use
Full data migration mode or to reprotect data after a failure.
Alternatively, you can place a host in the maintenance mode by using ESXCLI. Before placing a host in this mode,
ensure that you powered off the VMs that run on the host.
To perform an action before entering maintenance mode, run the following command on the host:
esxcli system maintenanceMode set --enable 1 --vsanmode=<str>
You can track the progress of data migration in the cluster. For more information see vSAN Monitoring and
Troubleshooting.
223
VMware vSAN 8.0
to reprotect data after a failure and the inability to use the Full data migration mode. For information about designing and
sizing fault domains, see "Designing and Sizing vSAN Fault Domains" in vSAN Planning and Deployment.
Consider a scenario where you have a vSAN cluster with 16 hosts. The hosts are spread across four racks, that is, four
hosts per rack. To tolerate an entire rack failure, create a fault domain for each rack. You can configure a cluster of such
capacity with the Failures to tolerate set to 1. If you want the Failures to tolerate set to 2, configure five fault domains in
the cluster.
When a rack fails, all resources including the CPU, memory in the rack become unavailable to the cluster. To reduce
the impact of a potential rack failure, configure fault domains of smaller sizes. Increasing the number of fault domains
increases the total amount of resource availability in the cluster after a rack failure.
When working with fault domains, follow these best practices.
• Configure a minimum of three fault domains in the vSAN cluster. For best results, configure four or more fault domains.
• A host not included in any fault domain is considered to reside in its own single-host fault domain.
• You do not need to assign every vSAN host to a fault domain. If you decide to use fault domains to protect the vSAN
environment, consider creating equal sized fault domains.
• When moved to another cluster, vSAN hosts retain their fault domain assignments.
• When designing a fault domain, place a uniform number of hosts in each fault domain.
For guidelines about designing fault domains, see "Designing and Sizing vSAN Fault Domains" in vSAN Planning and
Deployment.
• You can add any number of hosts to a fault domain. Each fault domain must contain at least one host.
224
VMware vSAN 8.0
The selected host is no longer part of the fault domain. Any host that is not part of a fault domain is considered to reside in
its own single-host fault domain.
You can add hosts to fault domains. See Move Host into Selected Fault Domain in vSAN Cluster.
225
VMware vSAN 8.0
All hosts in the fault domain are removed and the selected fault domain is deleted from the vSAN cluster. Each host that is
not part of a fault domain is considered to reside in its own single-host fault domain.
226
VMware vSAN 8.0
vSAN data protection requires the VMware Snapshot Service to manage vSAN snapshots. Deploy the Snapshot Service
appliance to enable vSAN data protection in the vSphere Client.
Use the following tabs to navigate the vSAN data protection page.
Tab Description
Summary Displays general information about vSAN data protection, including the number of protection groups,
percentage of protected VMs, number of VM snapshots, and amount of storage space used for snapshots.
Protection Groups Displays a list of vSAN data protection groups and their status. Select a protection group to view snapshots
in the protection group, or edit the configuration.
VMs Displays a list of VMs in the vSAN cluster with details about their data protection status. Deleted VMs that
have snapshots available are visible here.
You can select a VM and click to restore or clone the VM.
227
VMware vSAN 8.0
vSAN Snapshots
vSAN snapshots preserve the state and data of a virtual machine at the time you take the snapshot. This local archive
preserves the VM's data as it existed at that time. You can restore a VM to the state that existed when the snapshot was
taken, or create a linked clone VM that matches the state preserved in the snapshot.
Taking a snapshot captures the VM state at a specific point in time. vSAN snapshots are not quiesced, and they capture
the current running state of the VM.
Snapshots operate on individual virtual machines. Each VM requires a separate snapshot. You can take manual or
scheduled snapshots of virtual machines by placing them in protection groups.
Each vSAN snapshot contains the state of the VM's namespace object and virtual disk objects. vSAN take snapshots of
VMs in protection groups at scheduled intervals. These vSAN snapshots are stored locally in the vSAN datastore.
Protection Groups
Protection groups enable you to schedule and manage snapshots for one or multiple VMs. You can add VMs to a
protection group, configure snapshot schedules, and view snapshot information.
Select a protection group, and use the following tabs to manage the group.
Tab Description
Overview Displays general information about the protection group, including a list of member VMs, the snapshot
schedules, and the number of snapshots taken.
Snapshots Displays the snapshot series associated with the protection group. You can select and delete individual
snapshots from the series.
VMs Displays a list of VMs that are members of the protection group, and the number of snapshots available for
each VM.
When you create a protection group, add member VMs and configure one or more snapshot schedules. You can add VMs
individually, or enter VM name patterns to add all VMs that match the pattern. You can use both methods to add VMs to
the protection group.
You can define multiple snapshot schedules to periodically capture the state of VMs in a protection group. As new
snapshots are captured, vSAN removes old snapshots from the series, based on the retention setting. You also can take a
manual snapshot to capture the current state of VMs in the protection group.
Enable immutability mode on a protection group for additional security. You cannot edit or delete this protection group,
change the VM membership, edit or delete snapshots. An immutable snapshot is a read-only copy of data that cannot be
modified or deleted, even by an attacker with administrative privileges.
NOTE
Once immutability mode is enabled on a protection group, it cannot be disabled by an administrator.
You can monitor and modify protection groups from the Protection Groups tab. Click a protection group to view details.
• Overview displays general information about the protection group, including VM membership, snapshot schedules,
and number of snapshots.
• Snapshots displays a list of snapshots available in the protection group. You can select a snapshot, and click >> to
view individual snapshots for each VM, and perform actions.
• VMs displays a list of VMs in the protection group with details about the available snapshots. Select a VM radio button,
and click Restore VM or Clone VM, then choose a snapshot.
228
VMware vSAN 8.0
Click one of the following buttons to perform actions on the protection group.
Action Description
Take snapshot You can change the default name of the snapshot, and define the retention period. vSAN takes a separate
snapshot for each VM in the protection group.
Edit You can add or remove VMs, modify the VM name patterns, and add or modify the snapshot schedules.
Pause schedule/ You can pause the snapshot schedules defined for the protection group. No snapshots are taken or deleted
Resume schedule while the schedules are paused.
To delete a protection group, click the More... icon next to the group name, and select menu Delete. When you delete the
protection group, you must decide how to manage its snapshots.
• Keep snapshots until their expiration date. The protection group will be deleted after all existing snapshots have
expired.
• Delete snapshots. The protection group and its existing snapshots are deleted immediately.
229
VMware vSAN 8.0
1. Download the VMware Snapshot Service appliance from the Broadcom website at https://support.broadcom.com/
group/ecx/downloads.
2. Right-click the vSAN cluster in the vSphere Client, and select Deploy OVF Template to open the wizard.
3. On the Select an OVF template page, specify the location of the appliance OVA file and click Next.
4. On the Select a name and folder page, you can enter a unique name for the appliance, select your data center as the
deployment location.
5. On the Select a compute resource page, select the vSAN cluster as the compute resource.
6. On the Select storage page, select a datastore.
7. On the Select networks page, select the same network as the vCenter, and click Next.
8. On the Customize template page, enter FQDN in the vCenter Server Hostname field. For example, sfo-w01-
dp01.sfo.rainpole.io. The vCenter server credentials are used only once to create a dedicated local service account.
The VMware Snapshot Service is deployed to the specified vCenter, and vSAN data protection pages are available in the
vSphere Client.
230
VMware vSAN 8.0
231
VMware vSAN 8.0
1. Navigate to the vSAN cluster in the vSphere Client, and select Configure > vSAN > Data Protection.
2. Select the Protection Groups tab, click a protection group, and select the Snapshots tab.
3. Select a snapshot, and click Delete Snapshot.
4. Click Delete.
232
VMware vSAN 8.0
For more information about using the vSAN iSCSI target service, refer to the iSCSI Target Usage Guidehttps://
core.vmware.com/resource/vsan-iscsi-target-usage-guide.
233
VMware vSAN 8.0
iSCSI Targets
You can add one or more iSCSI targets that provide storage blocks as logical unit numbers (LUNs). vSAN identifies each
iSCSI target by a unique iSCSI qualified Name (IQN). You can use the IQN to present the iSCSI target to a remote iSCSI
initiator so that the initiator can access the LUN of the target.
Each iSCSI target contains one or more LUNs. You define the size of each LUN, assign a vSAN storage policy to each
LUN, and enable the iSCSI target service on a vSAN cluster. You can configure a storage policy to use as the default
policy for the home object of the vSAN iSCSI target service.
234
VMware vSAN 8.0
changes back to the original I/O owner location as per the configuration. You can select one of the following
options to set the site location:
• Either: Hosts the iSCSI target service either on Preferred or Secondary site.
• Preferred: Hosts the iSCSI target service on the Preferred site.
• Secondary: Hosts the iSCSI target service on the Secondary site.
3. Click OK.
iSCSI target is created and listed under the vSAN iSCSI Targets section with the information such as IQN, I/O owner host,
and so on.
Define a list of iSCSI initiators that can access this target.
235
VMware vSAN 8.0
until they are added to the initiator group. You must check the current initiator connections and ensure that all
the authorized initiators are added to the initiator group before group creation.
1. Navigate to the vSAN cluster.
2. Click the Configure tab.
a) Under vSAN, click iSCSI Target Service.
b) Click the Initiator Groups tab, and click Add. The New Initiator Group dialog box is displayed.
c) Enter a name for the iSCSI initiator group.
d) (Optional) To add members to the initiator group, enter the IQN of each member. Use the following format to enter
the member IQN:
iqn.YYYY-MM.domain:name
Where:
• YYYY = year, such as 2016
• MM = month, such as 09
• domain = domain where the initiator resides
• name = member name (optional)
3. Click OK or Create.
Add members to the iSCSI initiator group.
236
VMware vSAN 8.0
3. Click the Enable vSAN iSCSI Target Service slider to turn it off and click Apply.
237
VMware vSAN 8.0
When you configure vSAN file service, vSAN creates a single VDFS distributed file system for the cluster which will be
used internally for management purposes. A file service VM (FSVM) is placed on each host. The FSVMs manage file
shares in the vSAN datastore. Each FSVM contains a file server that provides both NFS and SMB service.
A static IP address pool should be provided as an input while enabling file service workflow. One of the IP addresses is
designated as the primary IP address. The primary IP address can be used for accessing all the shares in the file services
cluster with the help of SMB and NFSv4.1 referrals. A file server is started for every IP address provided in the IP pool.
A file share is exported by only one file server. However, the file shares are evenly distributed across all the file servers.
To provide computing resources that help manage access requests, the number of IP addresses must be equal to the
number of hosts in the vSAN cluster.
vSAN file service supports vSAN stretched clusters and two-node vSAN clusters. A two-node vSAN cluster should have
two data node servers in the same location or office, and the witness in a remote or shared location.
For more information about Cloud Native Storage (CNS) file volumes, see the VMware vSphere Container Storage Plug-in
documentation and vSphere with Tanzu Configuration and Management documentation.
238
VMware vSAN 8.0
• vSAN 8.0 Update 3 ESA cluster supports 250 file shares. Out of those 250 file shares, maximum 100 file shares can
be SMB. For example, if you create 100 SMB file shares then the cluster can only support additional 150 NFS file
shares.
• vSAN File Services does not support the following:
– Read-Only Domain Controllers (RODC) for joining domains because the RODC cannot create machine accounts.
As a security best practice, a dedicated org unit should be pre-created in the Active Directory and the user name
mentioned here should be controlling this organization.
– Disjoint namespace.
– Multi domain and Single Active Directory Forest environments.
• When a host enters maintenance mode, the file server moves to another FSVM. The FSVM on the host that entered
maintenance mode is powered off. After the host exits maintenance mode, the FSVM is powered on.
• vSAN File Services VM (FSVM) docker internal network may overlap with the customer network without warning or
reconfiguration.
There is known conflict issue if the specified file service network overlaps with the docker internal network
(172.17.0.0/16). This causes routing problem for the traffic to the correct endpoint.
As a workaround, specify a different file service network so that it does not overlap with the docker internal network
(172.17.0.0/16).
239
VMware vSAN 8.0
Ensure that the following are configured before enabling the vSAN File Services:
• The vSAN cluster must be a regular vSAN cluster, a vSAN stretched cluster, or a vSAN ROBO cluster.
• Every ESXi host in the vSAN cluster must have minimal hardware requirements such as:
– 4 Core CPU
– 16 GB physical memory
• You must ensure to prepare the network as vSAN File Service network:
– If using standard switch based network, the Promiscuous Mode and Forged Transmits are enabled as part of the
vSAN File Services enablement process.
– If using DVS based network, vSAN File Services are supported on DVS version 6.6.0 or later. Create a dedicated
port group for vSAN File Services in the DVS. MacLearning and Forged Transmits are enabled as part of the vSAN
File Services enablement process for a provided DVS port group.
– IMPORTANT
If using NSX-based network, ensure that MacLearning is enabled for the provided network entity from
the NSX admin console, and all the hosts and File Services nodes are connected to the desired NSX-T
network.
1. Navigate to the vSAN cluster and click Configure > vSAN > Services.
2. On the File Service row, click Enable.
The Enable File Service wizard
240
opens.
VMware vSAN 8.0
Option Description
Automatically load latest OVF This option lets the system search and download the OVF.
NOTE
• Ensure that you have configured the proxy and
firewall so that vCenter can access the following
website and download the appropriate JSON file.
https://download3.vmware.com/software/
VSANOVF/FsOvfMapping.json
For more information about configuring the vCenter
DNS, IP address, and proxy settings, see vCenter
Server Appliance Configuration.
• Use current OVF: Lets you use the OVF that is
already available.
• Automatically load latest OVF: Lets the system
search and download the latest OVF.
Manually load OVF This option allows you to browse and select an OVF that is
already available on your local system.
NOTE
If you select this option, you should upload all the
following files:
• VMware-vSAN-File-Services-Applianc
e-x.x.x.x-x_OVF10.mf
• VMware-vSAN-File-Services-Applianc
e-x.x.x.x-x-x_OVF10.cert
• VMware-vSAN-File-Services-Applianc
e-x.x.x.x-x-x-system.vmdk
• VMware-vSAN-File-Services-Appliance-x.x.x.x-x-cl
oud-components.vmdk
• VMware-vSAN-File-Services-Appliance-x.x.x.x-x-lo
g.vmdk
• VMware-vSAN-File-Services-Appliance-x.x.x.x-x_O
VF10.ovf
5. Click Enable.
241
VMware vSAN 8.0
242
VMware vSAN 8.0
3. In the File service domain page, enter the unique namespace and click Next. The domain name must have minimum
two characters. The first character must be an alphabet or a number. The remaining characters can include an
alphabet, a number, an underscore ( _ ), a period ( . ), a hyphen ( - ).
4. In the Networking page, enter the following information, and click Next:
• Protocol: You can select IPv4 or IPv6. vSAN File Service only supports IPv4 or IPv6 stack. The reconfiguration
between IPv4 and IPv6 is not supported.
• DNS servers: Enter a valid DNS server to ensure the proper configuration of File Service.
• DNS suffixes: Provide the DNS suffix that is used with the file service. All other DNS suffixes from where the
clients can access these file servers must also be included. File Service does not support DNS domain with
single label, such as "app", "wiz", "com" and so on. A domain name given to file service must be of the format
thisdomain.registerdrootdnsname. DNS name and suffix must adhere to the best practices detailed in https://
docs.microsoft.com/en-us/windows-server/identity/ad-ds/plan/selecting-the-forest-root-domain.
• Subnet mask: Enter a valid subnet mask. This text box appears when you select IPv4.
• Prefix length: Enter a number between 1 and 128. This text box appears when you select IPv6.
• Gateway: Enter a valid gateway.
• IP Pool: Enter primary IP address and DNS name.
With vSAN 8.0 Update 3, vSAN ESA cluster supports 250 file shares. Out of those 250 file shares, maximum 100 file
shares can be SMB. For example, if you create 100 SMB file shares then the cluster can only support additional 150
NFS file shares.
Each file server on a vSAN ESA cluster can support a maximum of 25 file shares and requires at least 10 IPs to have
the maximum of 250 shares. With the increase in the file servers or file shares per host, there might be an impact
on the performance of vSAN File Service. For best performance, the number of IP address must to be equal to the
number of hosts in the vSAN cluster.
Affinity site option is available if you are configuring vSAN file service on a vSAN stretched cluster. This option allows
you to configure the placement of the file server on Preferred or Secondary site. This helps in reducing the cross-site
traffic latency. The default value is Either, which indicates that no site affinity rule is applied to the file server.
NOTE
If your cluster is a ROBO cluster, ensure that the Affinity site value is set to Either.
In a site failure event, the file server affiliated to that site fails over to the other site. The file server fails back to the
affiliated site when it is recovered. Configure more file servers to one site if more workloads can be expected from a
certain site.
NOTE
If the file server contains SMB file shares, then it does not failback automatically even if the site failure is
recovered.
Consider the following while configuring the IP addresses and DNS names:
• To ensure proper configuration of File Service, the IP addresses you enter in the Networking page must be static
addresses and the DNS server must have records for those IP addresses. For best performance, the number of IP
addresses must be equal to the number of hosts in the vSAN cluster.
• You can have a maximum of 64 hosts in the cluster. If large scale cluster support is configured, you can enter up to
64 IP addresses.
• You can use the following options to automatically fill the IP address and DNS server name text boxes:
AUTO FILL: This option is displayed after you enter the first IP address in the IP address text box. Click the AUTO
FILL option to automatically fill the remaining fields with sequential IP addresses, based on the subnet mask and
gateway address of the IP address that you have provided in the first row. You can edit the auto filled IP addresses.
LOOK UP DNS: This option is displayed after you enter the first IP address in the IP address text box. Click the
LOOK UP DNS option to automatically retrieve the FQDN corresponding to the IP addresses in the IP address
column.
243
VMware vSAN 8.0
NOTE
• All valid rules apply for the FQDNs. For more information, see https://tools.ietf.org/html/rfc953.
• The first part of the FQDN, also known as NetBIOS Name, must not have more than 15 characters.
The FQDNs are automatically retrieved only under the following conditions:
– You must have entered a valid DNS server in the Domain page.
– The IP addresses entered in the IP Pool page must be static addresses and the DNS server must have records
for those IP addresses.
5. In the Directory service page, enter the following information and click Next.
Option Description
Directory service Configure an Active Directory domain to vSAN File Service for
authentication. If you are planning to create an SMB file share or
an NFSv4.1 file share with Kerberos authentication, then you must
configure an AD domain to vSAN File Service.
AD domain Fully qualified domain name joined by the file server.
Preferred AD Server Enter the IP address of the preferred AD server. In case of multiple
IP addresses, ensure that they are separated by comma.
Organizational unit (Optional) Contains the computer account that the vSAN File Service
creates. In an organization with complex hierarchies, create the
computer account in a specified container by using a forward
slash mark to denote hierarchies (for example, organizational_unit/
inner_organizational_unit).
NOTE
By default, the vSAN File Service create the computer
account in the Computers container.
AD username User name to be used for connecting and configuring the Active
Directory service.
This user name authenticates the active directory on the domain.
A domain user authenticates the domain controller and creates
vSAN File Service computer accounts, related SPN entries, and
DNS entries (when using Microsoft DNS). As a best practice,
create a dedicated service account for the file service.
A domain user in the directory service with the following sufficient
privileges to create and delete computer objects:
• (Optional) Add/Update DNS entries
244
VMware vSAN 8.0
Option Description
Password Password for the user name of the Active Directory on the domain.
vSAN File Service use the password to authenticate to AD and to
create the vSAN File Service computer account.
NOTE
• vSAN File Service does not support the following:
– Read-Only Domain Controllers (RODC) for joining domains because the RODC cannot create
machine accounts. As a security best practice, a dedicated org unit must be pre-created in the Active
Directory and the user name mentioned here must be controlling this organization.
– Disjoint namespace.
– Multi domain and Single Active Directory Forest environments.
• Only English characters are supported for Active Directory user name.
• Only single AD domain configuration is supported. However, the file servers can be put on a valid DNS
subdomain. For example, an AD domain with the name example.com can have file server FQDN as
name1.eng.example.com .
• Pre-created computer objects for file servers are not supported. Make sure that the user provided here
have sufficient privilege over the organizational unit.
• vSAN File Service update the DNS records for the file servers if the Active Directory is also used as a
DNS server and the user has sufficient permission to update the DNS records. vSAN File Service also
has a Health Check to indicate if the forward and reverse lookups for file servers are working properly.
However, if there are other proprietary solutions used as DNS servers, the Vi admin must update these
DNS records.
6. Review the settings and click Finish.
The file service domain is configured. File servers are started with the IP addresses that were assigned during the vSAN
File Service configuration process.
1. Navigate to the vSAN cluster and click Configure > vSAN > Services.
2. On the File Service row, click Edit> Edit domain.
The File Service Domain wizard opens.
245
VMware vSAN 8.0
3. In the File service domain page, edit the file service domain name and click Next.
4. In the Networking page, make the appropriate configuration changes and click Next. You can edit the primary IP
addresses, static IP addresses, and DNS names. You can add or remove the primary IP addresses or static IP
addresses. You cannot change the DNS name without changing the IP.
NOTE
Changing domain information is a disruptive action. It might require all clients to use new URLs to reconnect
to the file shares.
5. In the Directory service page, make appropriate directory related changes and click Next.
NOTE
You cannot change the AD domain, organizational unit, and username after initially configuring vSAN File
Services.
6. In the Review page, click Finish after making necessary changes.
246
VMware vSAN 8.0
If you select the SMB protocol, you can also configure the SMB file share to accept only the encrypted data using
the Protocol encryption option.
If you select the NFS protocol, you can configure the file share to support either NFS 3, NFS 4, or both NFS 3 and
NFS 4 versions. If you select NFS 4 version, you can set either AUTH_SYS or Kerberos security.
NOTE
SMB protocol and Kerberos security for NFS protocol can be configured only if the vSAN File Service is
configured with Active Directory. For more information, see Configure vSAN File Service.
• With SMB protocol, you can hide the files and folders that the share client user does not have permission to access
using the Access based enumeration option.
• Storage Policy: Select an appropriate storage policy.
• Affinity site: This option is available if you are creating a file share on a vSAN stretched cluster. This option helps
you place the file share on a file server that belongs to the site of your choice. Use this option when you prefer low
latency while accessing the file share. The default value is Either, which indicates that the file share is placed on a
site with less traffic on either preferred or secondary site.
• Storage space quotas: You can set the following values:
– Share warning threshold: When the share reaches this threshold, a warning message is displayed.
– Share hard quota: When the share reaches this threshold, new block allocation is denied.
• Labels: A label is a key-value pair that helps you organize file shares. You can attach labels to each file share and
then filter them based on their labels. A label key is a string with 1~250 characters. A label value is a string and the
length of the label value should be less than 1k characters. vSAN File Service supports up to 5 labels per share.
4. The Net access control page, provides options to define access to the file share. Net access control options are
available only for NFS shares. Select one of the following options and click Next.
• No access: Select this option to make the file share inaccessible from any IP address.
• Allow access from any IP: Select this option to make the file share accessible from all IP addresses.
• Customize net access: Select this option to define permissions for specific IP addresses. Using this option you
can specify whether a particular IP address can access, make changes, or only read the file share. You can also
enable Root squash for each IP address. You can enter the IP addresses in the following formats:
– A single IP address. For example, 123.23.23.123
– IP address along with a subnet mask. For example, 123.23.23.0/8
– A range by specifying a starting IP address and ending IP address separated by a hyphen ( - ). For example,
123.23.23.123-123.23.23.128
– Asterisk ( * ) to imply all the clients.
5. In the Review page, review the settings, and then click Finish.
A new file share is created on the vSAN datastore.
247
VMware vSAN 8.0
You can access a file share from a host client, using an operating system that communicates with NFS file systems.
For RHEL-based Linux distributions, NFS 4.1 support is available in RHEL 7.3 and CentOS 7.3-1611 running kernel
3.10.0-514 or later. For Debian based Linux distributions, NFS 4.1 support is available in Linux kernel version 4.0.0 or
later. All NFS clients must have unique hostnames for NFSv4.1 to work. You can use the Linux mount command with the
Primary IP to mount a vSAN file share to the client. For example: mount -t nfs4 -o minorversion=1,sec=sys
<primary ip>:/vsanfs/<share name>. NFSv3 support is available for RHEL-based and Debian based Linux
distributions. You can use the Linux mount command to mount a vSAN file share to the client. For example: mount -t
nfs vers=3 <nfsv3_access_point> <localmount_point>.
Sample v41 commands for verifying the NFS file share from a host client:
[root@localhost ~]# mount -t nfs4 -o minorversion=1,sec=sys <primary ip address>:/vsan-
fs/TestShare-0 /mnt/TestShare-0
[root@localhost ~]# cd /mnt/TestShare-0/
[root@localhost TestShare-0]# mkdir bar
[root@localhost TestShare-0]# touch foo
[root@localhost TestShare-0]# ls -l
total 0
drwxr-xr-x. 1 root root 0 Feb 19 18:35 bar
-rw-r--r--. 1 root root 0 Feb 19 18:35 foo
Access NFS Kerberos File Share
A Linux client accessing an NFS Kerberos share should have a valid Kerberos ticket.
Sample v41 commands for verifying the NFS Kerberos file share from a host client: An NFS Kerberos
share can be mounted using the following mount command:
[root@localhost ~]# mount -t nfs4 -o minorversion=1,sec=krb5/krb5i/krb5p <primary ip ad-
dress>:/vsanfs/TestShare-0 /mnt/TestShare-0
[root@localhost ~]# cd /mnt/TestShare-0/
[root@localhost TestShare-0]# mkdir bar
[root@localhost TestShare-0]# touch foo
[root@localhost TestShare-0]# ls -l
total 0
drwxr-xr-x. 1 root root 0 Feb 19 18:35 bar
-rw-r--r--. 1 root root 0 Feb 19 18:35 foo
Changing Ownership of a NFS Kerberos share You must log in using the AD domain user name for
changing the ownership of a share. The AD domain user name provided in the file service configuration acts
as a sudo user for the Kerberos file share.
[root@localhost ~]# mount -t nfs4 -o minorversion=1,sec=sys <primary ip address>:/vsan-
fs/TestShare-0 /mnt/TestShare-0
[fsadmin@ocalhost ~]# chown user1 /mnt/TestShare-0
[user1@localhost ~]# ls -l /mnt/TestShare-0
total 0
drwxr-xr-x. 1 user1 domain users 0 Feb 19 18:35 bar
-rw-r--r--. 1 user1 domain users 0 Feb 19 18:35 foo
248
VMware vSAN 8.0
Ensure that the Windows client is joined to the Active Directory domain that is configured with vSAN File Service.
1. Copy the SMB file share path using the following procedure:
1. Navigate to the vSAN cluster and click Configure > vSAN > File Service Shares.
List of all the vSAN file shares appears.
2. Select the SMB file share that you want to access from the Windows client.
3. Click COPY PATH > SMB.
The SMB file share path gets copied to your clipboard.
2. Log into the Windows client as a normal Active Directory domain user.
3. Access the SMB file share using path that you have copied.
249
VMware vSAN 8.0
2. Log into the Windows client as a file service admin user. The file service admin user is configured when you create the
file service domain. A file service admin user has all the privileges on the file server.
3. In the search box on the taskbar, type Run, and then select Run.
4. In the Run box, run the MMC command that you have copied to access and manage the SMB share using the MMC
tool.
250
VMware vSAN 8.0
– ~ (ASCII 126)
Create a Snapshot
When the vSAN file service is enabled, you can create one or more snapshots that provide a point-in-time image of the
vSAN file share. You can create a maximum of 32 snapshots per file share.
You should have created a vSAN file share.
1. Navigate to the vSAN cluster and click Configure > vSAN > File Service Shares.
A list of vSAN file shares appears.
2. Select the file share for which you want to create a snapshot and then click SNAPSHOTS > NEW SNAPSHOT.
Create new snapshot dialogue appears.
3. On the Create new snapshot dialogue, provide a name for the snapshot, and click Create.
You can view the list of snapshots along with the information such as date and time of the snapshot creation, and its size.
1. Navigate to the vSAN cluster and click Configure > vSAN > File Service Shares.
A list of vSAN file shares appears.
2. Select a file share and click SNAPSHOTS.
A list of snapshots for that file share appears. You can view information such as date and time of the snapshot creation,
and its size.
Delete a Snapshot
If there is an imbalance in the workload of a host, you can correct it by rebalancing the workload.
1. Navigate to the vSAN cluster and then click Monitor > vSAN > Skyline Health.
2. Under Skyline Health, expand File Service and then click Infrastructure Health.
The Infrastructure Health tab displays a list of all the hosts that are part of the vSAN File Service infrastructure. For
each host, the status of workload balance is displayed. If there is an imbalance in the workload of a host, an alert is
displayed in the Description column.
251
VMware vSAN 8.0
The host workload is balanced and the workload balance status turns green.
When you enable unmap on a vSAN cluster, you must power off and then power on all VMs. VMs must use virtual
hardware version 13 or above to perform unmap operations.
252
VMware vSAN 8.0
Option Action
NOTE
If you select this option, you should upload all the
following files:
• VMware-vSAN-File-Services-
Appliance-x.x.x.x-x_OVF10.mf
• VMware-vSAN-File-Services-
Appliance-x.x.x.x-x-x_OVF10.cert
• VMware-vSAN-File-Services-
Appliance-x.x.x.x-x-x-system.vmdk
• VMware-vSAN-File-Services-
Appliance-x.x.x.x-x-cloud-
components.vmdk
• VMware-vSAN-File-Services-
Appliance-x.x.x.x-x-log.vmdk
• VMware-vSAN-File-Services-
Appliance-x.x.x.x-x_OVF10.ovf
The throughput, IOPS, and latency metrics of the vSAN file service for the selected period are displayed.
For more information on vSAN Performance Graphs, see the VMware knowledge base article at https://kb.vmware.com/s/
article/2144493.
253
VMware vSAN 8.0
254
VMware vSAN 8.0
NOTE
Follow the steps to migrate a hybrid vSAN cluster to Solid State Drive (SSD), hybrid vSAN cluster to NVMe, or
SSD to NVMe.
1. Remove the hybrid disk groups on the host.
a) In the vSphere Client, navigate to the vSAN cluster, and click the Configure tab.
b) Under vSAN, click Disk Management.
c) Under Disk Groups, select the disk group to remove, click …, and then click Remove.
Select Full data migration as a migration mode and click Yes.
NOTE
Migrate the disk groups on each host in the vSAN cluster.
2. Remove the physical HDD disks from the host.
3. Add the flash devices to the host.
Verify that no partitions exist on the flash devices.
4. Create the all-flash disk groups on the host.
5. Repeat the steps 1 through 4 on each host until all the hybrid disk groups are converted to the all-flash disk groups.
NOTE
If you cannot hot-plug disks on the host, place the host in maintenance mode before removing disks in the
vSphere Client. Shut down the host to replace the disks with flash devices. Then power on the host, exit
maintenance mode, and create new disk groups.
255
VMware vSAN 8.0
Shut Down the vSAN Cluster Using the Shutdown Cluster Wizard
Use the Shutdown cluster wizard to gracefully shut down the vSAN cluster for maintenance or troubleshooting.
The Shutdown Cluster Wizard is available with vSAN 7.0 Update 3 and later releases.
NOTE
If you have a vSphere with Tanzu environment, you must follow the specified order when shutting down or
starting up the components. For more information, see "Shutdown and Startup of VMware Cloud Foundation" in
the VMware Cloud Foundation Operations Guide.
1. Prepare the vSAN cluster for shutdown.
a) Check the vSAN Skyline Health to confirm that the cluster is healthy.
b) Power off all virtual machines (VMs) stored in the vSAN cluster, except for vCenter Server VMs, vCLS VMs and
file service VMs. If vCenter Server is hosted on the vSAN cluster, do not power off the vCenter Server VM or VM
service VMs (such as DNS, Active Directory) used by vCenter Server.
c) If this is an HCI Mesh server cluster, power off all client VMs stored on the cluster. If the client cluster's vCenter
Server VM is stored on this cluster, either migrate or power off the VM. Once this server cluster is shutdown, its
shared datastore is inaccessible to clients.
d) Verify that all resynchronization tasks are complete.
Click the Monitor tab and select vSAN > Resyncing Objects.
NOTE
If any member hosts are in lockdown mode, add the host's root account to the security profile Exception User
list. For more information, see Lockdown Mode in vSphere Security.
2. Right-click the vSAN cluster in the vSphere Client, and select menu Shutdown cluster.
You also can click Shutdown Cluster on the vSAN Services page.
3. On the Shutdown cluster wizard, verify that the Shutdown pre-checks are green checks. Resolve any issues that are
red exclamations. Click Next.
If vCenter Server appliance is deployed on the vSAN cluster, the Shutdown cluster wizard displays the vCenter Server
notice. Note the IP address of the orchestration host, in case you need it during the cluster restart. If vCenter Server
uses service VMs such as DNS or Active Directory, note them as exceptional VMs in the Shutdown cluster wizard.
4. Enter a reason for performing the shutdown, and click Shutdown.
The vSAN Services page changes to display information about the shutdown process.
5. Monitor the shutdown process.
vSAN performs the steps to shutdown the cluster, powers off the system VMs, and powers off the hosts.
Restart the vSAN cluster. See Restart the vSAN Cluster.
256
VMware vSAN 8.0
4. After the cluster has restarted, check the vSAN Skyline Health and resolve any outstanding issues.
257
VMware vSAN 8.0
j) Place all the hosts into maintenance mode with No Action. If the vCenter Server is powered off, use the following
command to place the ESXi hosts into maintenance mode with No Action.
esxcli system maintenanceMode set -e true -m noAction
258
VMware vSAN 8.0
3. If there are unhealthy or disconnected hosts in the cluster, recover or remove the hosts from the vSAN cluster. If
vCenter Server uses service VMs such as DNS or Active Directory, note them as exceptional VMs in the Shutdown
cluster wizard.
Retry the above commands only after the vSAN Skyline Health shows all available hosts in the green state.
If you have a three-node vSAN cluster, the command reboot_helper.py recover cannot work in a one host failure
situation. As an administrator, do the following:
1. Temporarily remove the failure host information from the unicast agent list.
2. Add the host after running the following command.
reboot_helper.py recover
Following are the commands to remove and add the host to a vSAN cluster:
#esxcli vsan cluster unicastagent remove -a <IP Address> -t node -u <NodeUuid>
#esxcli vsan cluster unicastagent add -t node -u <NodeUuid> -U true -a <IP Address> -p 12321
259
VMware vSAN 8.0
You can claim host-local devices for vSAN Direct, and use vSAN to manage and monitor those devices. On each local
device, vSAN Direct creates and independent VMFS datastore and makes it available to your stateful application.
Each local vSAN Direct datastore appears as a vSAN-D datastore.
NOTE
If vSAN Express Storage Architecture is enabled for the cluster, you cannot claim disks for vSAN Direct.
260
VMware vSAN 8.0
261
VMware vSAN 8.0
You can choose devices that are not reported as certified for vSAN ESA and those devices will be considered in the
storage pool, but such configuration is not recommended and can impact performance.
1. Navigate to the vSAN cluster.
2. Click the Configure tab.
3. Under vSAN, click Disk Management.
4. To manually claim disks, click Claim Unused Disks.
a) Select the devices you want to claim
b) Click Create.
5. To automatically claim disks, click CHANGE DISK CLAIM MODE and click the vSAN managed disk claim toggle
button.
NOTE
If you chose to use vSAN managed disk claiming when configuring the cluster, the toggle button would be
already enabled.
vSAN claims the devices that you selected and organizes them into storage pools that support the vSAN datastore.
By default, vSAN creates one storage pool for each ESXi host that contributes storage to the cluster. If the selected
devices are not certified for vSAN ESA, those devices are not considered for creating storage pools.
For each device you claim, vSAN creates a new vSAN Direct datastore.
You can click the Datastores tab to display the vSAN Direct datastores in your cluster.
262
VMware vSAN 8.0
Check a Disk or Disk Group's Data Migration Capabilities from vSAN Cluster
Use the data migration pre-check to find the impact of migration options when unmounting a disk or disk group, or
removing it from the vSAN cluster.
Run the data migration pre-check before you unmount or remove a disk or disk group from the vSAN cluster. The test
results provide information to help you determine the impact to cluster capacity, predicted health checks, and any objects
263
VMware vSAN 8.0
that will go out of compliance. If the operation will not succeed, pre-check provides information about what resources are
needed.
1. Navigate to the vSAN cluster.
2. Click the Monitor tab.
3. Under vSAN, click Data Migration Pre-check.
4. Select a disk or disk group, choose a data migration option, and click Pre-check.
vSAN runs the data migration precheck tests.
5. View the test results.
The pre-check results show whether you can safely unmount or remove the disk or disk group.
• The Object Compliance and Accessibility tab displays objects that might have issues after the data migration.
• The Cluster Capacity tab displays the impact of data migration on the vSAN cluster before and after you perform
the operation.
• The Predicted Health tab displays the health checks that might be affected by the data migration.
If the pre-check indicates that you can unmount or remove the device, click the option to continue the operation.
264
VMware vSAN 8.0
Option Description
2. Under Disks, select the device to remove, and click the
Remove Disk(s).
3. Select a data evacuation mode.
All data residing on the disks is evacuated. The disk group is removed from the cluster, and recreated.
Locator LEDs
You can turn locator LEDs on vSAN storage devices on or off. When you turn on the locator LED, you can identify the
location of a specific storage device.
• Verify that you have installed the supported drivers for storage I/O controllers that enable this feature. For information
about the drivers that are certified by VMware, see the VMware Compatibility Guide at http://www.vmware.com/
resources/compatibility/search.php.
• In some cases, you might need to use third-party utilities to configure the Locator LED feature on your storage I/O
controllers. For example, when you are using HP you should verify that the HP SSA CLI is installed.
265
VMware vSAN 8.0
For information about installing third-party VIBs, see the vSphere Upgrade documentation.
When you no longer need a visual alert on your vSAN devices, you can turn off locator LEDs on the selected devices.
1. Navigate to the vSAN cluster.
2. Click the Configure tab.
3. Under vSAN, click Disk Management.
4. Select a host to view the list of devices.
5. At the bottom of the page, select one or more storage devices from the list, and perform the desired action for the
locator LEDs.
Option Action
Turn on LED Turns on locator LED on the selected storage device. You also
can use the Manage tab and click Storage> Storage Devices.
Turn off LED Turns off locator LED on the selected storage device. You also
can use the Manage tab and click Storage> Storage Devices.
266
VMware vSAN 8.0
267
VMware vSAN 8.0
268
VMware vSAN 8.0
269
VMware vSAN 8.0
Reclaiming storage space can provide a higher host-to-flash I/O throughput and improve the flash endurance.
Unmap capability is not enabled by default. Enable Guest Trim/Unmap on the vSAN Services Advanced options tab.
When you enable unmap on a vSAN cluster, you must power off and then power on all VMs. VMs must use virtual
hardware version 13 or above to perform unmap operations.
vSAN also supports the SCSI UNMAP commands issued directly from a guest operating system to reclaim storage space.
vSAN supports offline unmaps and inline unmaps. On Linux OS, offline unmaps are performed with the fstrim(8)
command, and inline unmaps are performed when the mount -o discard command is used. On Windows OS, NTFS
performs inline unmaps by default.
270
VMware vSAN 8.0
Consider the following guidelines when managing disks in a cluster with deduplication and compression enabled. These
guidelines do not apply to compression-only vSAN.
• Avoid adding disks to a disk group incrementally. For more efficient deduplication and compression, consider adding a
disk group to increase the cluster storage capacity.
• When you add a disk group manually, add all the capacity disks at the same time.
• You cannot remove a single disk from a disk group. You must remove the entire disk group to make modifications.
• A single disk failure causes the entire disk group to fail.
You can view the Usage breakdown before dedup and compression when you monitor vSAN capacity in the vSphere
Client. It displays information about the results of deduplication and compression. The Used Before space indicates the
logical space required before applying deduplication and compression, while the Used After space indicates the physical
space used after applying deduplication and compression. The Used After space also displays an overview of the amount
of space saved, and the Deduplication and Compression ratio.
The Deduplication and Compression ratio is based on the logical (Used Before) space required to store data before
applying deduplication and compression, in relation to the physical (Used After) space required after applying
deduplication and compression. Specifically, the ratio is the Used Before space divided by the Used After space. For
example, if the Used Before space is 3 GB, but the physical Used After space is 1 GB, the deduplication and compression
ratio is 3x.
When deduplication and compression are enabled on the vSAN cluster, it might take several minutes for capacity updates
to be reflected in the Capacity monitor as disk space is reclaimed and reallocated.
271
VMware vSAN 8.0
272
VMware vSAN 8.0
While disabling deduplication and compression, vSAN changes the disk format on each disk group of the cluster.
It evacuates data from the disk group, removes the disk group, and recreates it with a format that does not support
deduplication and compression.
The time required for this operation depends on the number of hosts in the cluster and amount of data. You can monitor
the progress on the Tasks and Events tab.
273
VMware vSAN 8.0
maintaining full protection. Or a four-node cluster with RAID-5 objects already deployed. In the latter case, you have no
place to move part of the RAID-5 stripe, since RAID-5 objects require a minimum of four nodes.
You can still enable deduplication and compression and use the Allow Reduced Redundancy option. This option keeps
the VMs running, but the VMs might be unable to tolerate the full level of failures defined in the VM storage policy. As a
result, temporarily during the format change for deduplication and compression, your virtual machines might be at risk of
experiencing data loss. vSAN restores full compliance and redundancy after the format conversion is completed.
Table 28: Capacity Required to Store and Protect Data at Different RAID Levels
RAID 5 or RAID 6 erasure coding is a policy attribute that you can apply to virtual machine components. To use RAID 5,
set Failure tolerance method to RAID-5/6 (Erasure Coding) and Failures to tolerate to 1. To use RAID 6, set Failure
tolerance method to RAID-5/6 (Erasure Coding) and Failures to tolerate to 2. RAID 5 or RAID 6 erasure coding does
not support a Failures to tolerate value of 3.
274
VMware vSAN 8.0
To use RAID 1, set Failure tolerance method to RAID-1 (Mirroring). RAID 1 mirroring requires fewer I/O operations to
the storage devices, so it can provide better performance. For example, a cluster resynchronization takes less time to
complete with RAID 1.
NOTE
In a vSAN stretched cluster, the Failure tolerance method of RAID-5/6 (Erasure Coding) applies only to the
Site disaster tolerance setting.
NOTE
For a vSAN Express Storage Architecture cluster, depending on the number of fault domains that you use, the
number of components listed under RAID 5 (Monitor > vSAN> Virtual Objects> testVM > View Placement
Details) will vary. If six or more fault domains are available in the cluster, then five components will be listed
under RAID 5. If five or fewer fault domains are available, then three components will be listed.
For more information about configuring policies, see Using vSAN Policies.
275
VMware vSAN 8.0
vSAN data-in-transit encryption is a cluster-wide setting. When enabled, all data and metadata traffic is encrypted as it
transits across hosts.
Encryption of data in transit is enabled on the vSAN cluster. vSAN encrypts all data moving across hosts and file service
inter-host connections in the cluster.
vSAN Data-At-Rest Encryption
vSAN can encrypt data at rest in your vSAN datastore.
When you enable data at rest encryption, vSAN encrypts data after all other processing, such as deduplication, is
performed. Data at rest encryption protects data on storage devices, in case a device is removed from the cluster.
Using encryption on your vSAN datastore requires some preparation. After your environment is set up, you can enable
data-at-rest encryption on your vSAN cluster.
Data-at-rest encryption requires an external Key Management Server (KMS) or a vSphere Native Key Provider. For more
information about vSphere encryption, see vSphere Security.
You can use an external Key Management Server (KMS), the vCenter Server system, and your ESXi hosts to encrypt
data in your vSAN cluster. vCenter Server requests encryption keys from an external KMS. The KMS generates and
stores the keys, and vCenter Server obtains the key IDs from the KMS and distributes them to the ESXi hosts.
vCenter Server does not store the KMS keys, but keeps a list of key IDs.
276
VMware vSAN 8.0
When a host reboots, it does not mount its disk groups until it receives the KEK. This process can take several minutes
or longer to complete. You can monitor the status of the disk groups in the vSAN health service, under Physical disks >
Software state health.
For more information about encryption key persistence, see "Key Persistence Overview" in vSphere Security.
277
VMware vSAN 8.0
The password recrypts core dumps that use internal keys to use keys that are based on the password. You
can later use the password to decrypt any encrypted core dumps that might be included in the support bundle.
Unencrypted core dumps or logs are not affected.
– The password that you specify during vm-support bundle creation is not persisted in vSphere components. You are
responsible for keeping track of passwords for support bundles.
You add a Key Management Server (KMS) to your vCenter Server system from the vSphere Web Client.
• Verify that the key server is in the vSphere Compatibility Matrices and is KMIP 1.1 compliant.
– Verify that you have the required privileges:Cryptographer.ManageKeyServers
• Connecting to a KMS by using only an IPv6 address is not supported.
• Connecting to a KMS through a proxy server that requires user name or password is not supported.
vCenter Server creates a KMS cluster when you add the first KMS instance. If you configure the KMS cluster on two or
more vCenter Servers, make sure you use the same KMS cluster name.
NOTE
Do not deploy your KMS servers on the Virtual SAN cluster you plan to encrypt. If a failure occurs, hosts in the
Virtual SAN cluster must communicate with the KMS.
• When you add the KMS, you are prompted to set this cluster as a default. You can later change the default cluster
explicitly.
• After vCenter Server creates the first cluster, you can add KMS instances from the same vendor to the cluster.
• You can set up the cluster with only one KMS instance.
• If your environment supports KMS solutions from different vendors, you can add multiple KMS clusters.
1. Log in to the vCenter Server system with the vSphere Web Client.
2. Browse the inventory list and select the vCenter Server instance.
3. Click Configure and click Key Management Servers.
4. Click Add KMS, specify the KMS information in the wizard, and click OK.
Option Value
KMS cluster Select Create new cluster for a new cluster. If a cluster exists,
you can select that cluster.
Cluster name Name for the KMS cluster. You can use this name to connect to
the KMS if your vCenter Server instance becomes unavailable.
Server alias Alias for the KMS. You can use this alias to connect to the KMS
if your vCenter Server instance becomes unavailable.
Server address IP address or FQDN of the KMS.
Server port Port on which vCenter Server connects to the KMS.
278
VMware vSAN 8.0
Option Value
Proxy address Optional proxy address for connecting to the KMS.
Proxy port Optional proxy port for connecting to the KMS.
User name Some KMS vendors allow users to isolate encryption keys that
are used by different users or groups by specifying a user name
and password. Specify a user name only if your KMS supports
this functionality, and if you intend to use it.
Password Some KMS vendors allow users to isolate encryption keys that
are used by different users or groups by specifying a user name
and password. Specify a password only if your KMS supports
this functionality, and if you intend to use it.
Use the Root CA Certificate Option to Establish a Standard Key Provider Trusted Connection
Some Key Management Server (KMS) vendors require that you upload your root CA certificate to the KMS.
All certificates that are signed by your root CA are then trusted by this KMS. The root CA certificate that vSphere Virtual
Machine Encryption uses is a self-signed certificate that is stored in a separate store in the VMware Endpoint Certificate
Store (VECS) on the vCenter Server system.
279
VMware vSAN 8.0
NOTE
Generate a root CA certificate only if you want to replace existing certificates. If you do, other certificates that are
signed by that root CA become invalid. You can generate a new root CA certificate as part of this workflow.
1. Navigate to the .
2. Click Configure and select Key Providers under Security.
3. Select the key provider with which you want to establish a trusted connection.
The KMS for the key provider is displayed.
4. From the Establish Trust drop-down menu, select Make KMS trust vCenter.
5. Select vCenter Root CA Certificate and click Next.
The Download Root CA Certificate dialog box is populated with the root certificate that vCenter Server uses for
encryption. This certificate is stored in VECS.
6. Copy the certificate to the clipboard or download the certificate as a file.
7. Follow the instructions from your KMS vendor to upload the certificate to their system.
NOTE
Some KMS vendors require that the KMS vendor restarts the KMS to pick up the root certificate that you
upload.
Finalize the certificate exchange. See Finish the Trust Setup for a Standard Key Provider.
Use the Certificate Option to Establish a Standard Key Provider Trusted Connection
Some Key Management Server (KMS) vendors require that you upload the vCenter Server certificate to the KMS.
After the upload, the KMS accepts traffic that comes from a system with that certificate. vCenter Server generates a
certificate to protect connections with the KMS. The certificate is stored in a separate key store in the VMware Endpoint
Certificate Store (VECS) on the vCenter Server system.
1. Navigate to the .
2. Click Configure and select Key Providers under Security.
3. Select the key provider with which you want to establish a trusted connection.
The KMS for the key provider is displayed.
4. From the Establish Trust drop-down menu, select Make KMS trust vCenter.
5. Select vCenter Certificate and click Next.
The Download Certificate dialog box is populated with the root certificate that vCenter Server uses for encryption. This
certificate is stored in VECS.
NOTE
Do not generate a new certificate unless you want to replace existing certificates.
6. Copy the certificate to the clipboard or download it as a file.
7. Follow the instructions from your KMS vendor to upload the certificate to the KMS.
Finalize the trust relationship. See Finish the Trust Setup for a Standard Key Provider.
Use the New Certificate Signing Request Option to Establish a Standard Key Provider Trusted Connection
280
VMware vSAN 8.0
Some Key Management Server (KMS) vendors require that vCenter Server generate a Certificate Signing Request (CSR)
and send that CSR to the KMS.
The KMS signs the CSR and returns the signed certificate. You can upload the signed certificate to vCenter Server. Using
the New Certificate Signing Request option is a two-step process. First you generate the CSR and send it to the KMS
vendor. Then you upload the signed certificate that you receive from the KMS vendor to vCenter Server.
1. Navigate to the .
2. Click Configure and select Key Providers under Security.
3. Select the key provider with which you want to establish a trusted connection.
The KMS for the key provider is displayed.
4. From the Establish Trust drop-down menu, select Make KMS trust vCenter.
5. Select New Certificate Signing Request (CSR) and click Next.
6. In the dialog box, copy the full certificate in the text box to the clipboard or download it as a file.
Use the Generate new CSR button in the dialog box only if you explicitly want to generate a CSR.
7. Follow the instructions from your KMS vendor to submit the CSR.
8. When you receive the signed certificate from the KMS vendor, click Key Providers again, select the key provider, and
from the Establish Trust drop-down menu, select Upload Signed CSR Certificate.
9. Paste the signed certificate into the bottom text box or click Upload File and upload the file, and click Upload.
Finalize the trust relationship. See Finish the Trust Setup for a Standard Key Provider.
Use the Upload Certificate and Private Key Option to Establish a Standard Key Provider Trusted Connection
Some Key Management Server (KMS) vendors require that you upload the KMS server certificate and private key to the
vCenter Server system.
• Request a certificate and private key from the KMS vendor. The files are X509 files in PEM format.
Some KMS vendors generate a certificate and private key for the connection and make them available to you. After you
upload the files, the KMS trusts your vCenter Server instance.
1. Navigate to the .
2. Click Configure and select Key Providers under Security.
3. Select the key provider with which you want to establish a trusted connection.
The KMS for the key provider is displayed.
4. From the Establish Trust drop-down menu, select Make KMS trust vCenter.
5. Select KMS certificate and private key and click Next.
6. Paste the certificate that you received from the KMS vendor into the top text box or click Upload a File to upload the
certificate file.
7. Paste the key file into the bottom text box or click Upload a File to upload the key file.
8. Click Establish Trust.
Finalize the trust relationship. See Finish the Trust Setup for a Standard Key Provider.
Set the Default Key Provider Using the vSphere Client
You can use the vSphere Client to set the default key provider at the vCenter Server level.
281
VMware vSAN 8.0
As a best practice, verify that the Connection Status in the Key Providers tab shows Active and a green check mark.
You must set the default key provider if you do not make the first key provider the default, or if your environment uses
multiple key providers and you remove the default one.
1. Log in using the vSphere Client.
2. Navigate to the .
3. Click Configure and select Key Providers under Security.
4. Select the key provider.
5. Click Set as Default.
A confirmation dialog box appears.
6. Click Set as Default.
The key provider displays as the current default.
Finish the Trust Setup for a Standard Key Provider
Unless the Add Standard Key Provider dialog prompted you to trust the KMS, you must explicitly establish trust after
certificate exchange is complete.
You can complete the trust setup, that is, make vCenter Server trust the KMS, either by trusting the KMS or by uploading
a KMS certificate. You have two options:
• Trust the certificate explicitly by using the Upload KMS certificate option.
• Upload a KMS leaf certificate or the KMS CA certificate to vCenter Server by using the Make vCenter Trust KMS
option.
NOTE
If you upload the root CA certificate or the intermediate CA certificate, vCenter Server trusts all certificates that
are signed by that CA. For strong security, upload a leaf certificate or an intermediate CA certificate that the
KMS vendor controls.
1. Navigate to the .
2. Click Configure and select Key Providers under Security.
3. Select the key provider with which you want to establish a trusted connection.
The KMS for the key provider is displayed.
4. Select the KMS.
5. Select one of the following options from the Establish Trust drop-down menu.
Option Action
Make vCenter Trust KMS In the dialog box that appears, click Trust.
Upload KMS certificate 1. In the dialog box that appears, either paste in the certificate,
or click Upload a file and browse to the certificate file.
2. Click Upload.
282
VMware vSAN 8.0
Encryption of data at rest is enabled on the vSAN cluster. vSAN encrypts all data added to the vSAN datastore.
283
VMware vSAN 8.0
A rolling reformat of all disk groups takes places as vSAN encrypts all data in the vSAN datastore.
284
VMware vSAN 8.0
If data-at-rest encryption is enabled on a vSAN cluster, any core dumps in the vm-support package are encrypted.
Inform your support representative that data-at-rest encryption is enabled for the vSAN datastore. Your support
representative might ask you to decrypt core dumps to extract relevant information.
285
VMware vSAN 8.0
NOTE
Core dumps can contain sensitive information. Follow your organization's security and privacy policy to protect
sensitive information such as host keys.
You can collect the package, and you can specify a password if you expect to decrypt the core dump later. The vm-
support package includes log files, core dump files, and more.
1. Log in to vCenter Server using the vSphere Client.
2. Click Hosts and Clusters, and right-click the ESXi host.
3. Select Export System Logs.
4. In the dialog box, select Password for encrypted core dumps, and specify and confirm a password.
5. Leave the defaults for other options or make changes if requested by VMware Technical Support, and click Finish.
6. Specify a location for the file.
7. If your support representative asked you to decrypt the core dump in the vm-support package, log in to any ESXi host
and follow these steps.
a) Log in to the ESXi and connect to the directory where the vm-support package is located.
The filename follows the pattern esx.date_and_time.tgz.
b) Make sure that the directory has enough space for the package, the uncompressed package, and the
recompressed package, or move the package.
c) Extract the package to the local directory.
vm-support -x *.tgz .
The resulting file hierarchy might contain core dump files for the ESXi host, usually in /var/core, and might
contain multiple core dump files for virtual machines.
d) Decrypt each encrypted core dump file separately.
crypto-util envelope extract --offset 4096 --keyfile vm-support-incident-key-file
--password encryptedZdumpdecryptedZdump
vm-support-incident-key-file is the incident key file that you find at the top level in the directory.
encryptedZdump is the name of the encrypted core dump file.
decryptedZdump is the name for the file that the command generates. Make the name similar to the
encryptedZdump name.
e) Provide the password that you specified when you created the vm-support package.
f) Remove the encrypted core dumps, and compress the package again.
vm-support --reconstruct
You can decrypt or re-encrypt an encrypted core dump on your ESXi host by using the crypto-util CLI.
The ESXi host key that was used to encrypt the core dump must be available on the ESXi host that generated the core
dump.
You can decrypt and examine the core dumps in the vm-support package yourself. Core dumps might contain sensitive
information. Follow your organization's security and privacy policy to protect sensitive information, such as host keys.
For details about re-encrypting a core dump and other features of crypto-util, see the command-line help.
286
VMware vSAN 8.0
NOTE
crypto-util is for advanced users.
1. Log directly in to the ESXi host on which the core dump occurred.
If the ESXi host is in lockdown mode, or if SSH access is not enabled, you might have to enable access first.
2. Determine whether the core dump is encrypted.
Option Description
Monitor core dump crypto-util envelope describe vmmcores.ve
zdump file crypto-util envelope describe --offset 4096 zdump-
File
Upgrade Prerequisite
Consider the aspects that might delay the overall upgrade process. For guidelines and best practices, see the vSphere
Upgrade documentation.
Review the key requirements before you upgrade your cluster.
287
VMware vSAN 8.0
Software, hardware, drivers, firmware, and storage Verify that the new version of vSAN supports the software and hardware
I/O controllers components, drivers, firmware, and storage I/O controllers that you plan on using.
Supported items are listed on the VMware Compatibility Guide website at http://w
ww.vmware.com/resources/compatibility/search.php.
vSAN version Verify that you are using the latest version of vSAN. You cannot upgrade from a
beta version to the new vSAN. When you upgrade from a beta version, you must
perform a fresh deployment of vSAN.
Disk space Verify that you have enough space available to complete the software version
upgrade. The amount of disk storage needed for the vCenter Server installation
depends on your vCenter Server configuration. For guidelines about the disk
space required for upgrading vSphere, see the vSphere Upgrade documentation.
vSAN disk format vSAN disk format is a metadata upgrade that does not require data evacuation or
rebuilding.
vSAN hosts Verify that you have placed the vSAN hosts in maintenance mode and selected
the Ensure data accessibility or Evacuate all data option.
You can use the vSphere Lifecycle Manager for automating and testing the
upgrade process. However, when you use vSphere Lifecycle Manager to upgrade
vSAN, the default evacuation mode is Ensure data accessibility. When you
use the Ensure data accessibility mode, your data is not protected, and if you
encounter a failure while upgrading vSAN, you might experience unexpected data
loss. However, the Ensure data accessibility mode is faster than the Evacuate
all data mode, because you do not need to move all data to another host in the
cluster. For information about various evacuation modes, see the Administering
VMware vSAN documentation.
Virtual Machines Verify that you have backed up your virtual machines.
Recommendations
Consider the following recommendations when deploying ESXi hosts for use with vSAN:
• If ESXi hosts are configured with memory capacity of 512 GB or less, use SATADOM, SD, USB, or hard disk devices
as the installation media.
• If ESXi hosts are configured with memory capacity greater than 512 GB, use a separate magnetic disk or flash device
as the installation device. If you are using a separate device, verify that vSAN is not claiming the device.
• When you boot a vSAN host from a SATADOM device, you must use a single-level cell (SLC) device and the size of
the boot device must be at least 16 GB.
• To ensure your hardware meets the requirements for vSAN, refer to vSAN Planning and Deployment.
vSAN 6.5 and later enables you to adjust the boot size requirements for an ESXi host in a vSAN cluster.
288
VMware vSAN 8.0
Using vSphere Lifecycle Manager to upgrade hosts in parallel can result in the witness host being upgraded in parallel
with one of the data hosts. To avoid upgrade problems, configure vSphere Lifecycle Manager so it does not upgrade the
witness host in parallel with the data hosts.
289
VMware vSAN 8.0
Each vSAN release supports the on-disk format of prior releases. All hosts in the cluster must have the same on-disk
format version. Because some features are tied to the on-disk format version, it's best to upgrade the vSAN on-disk
format to the highest version supported by the ESXi version. For more information, refer to https://kb.vmware.com/s/
article/2148493.
vSAN on-disk format version 3 and higher require only a metadata upgrade that takes a few minutes. No disk evacuation
or reconfiguration is performed during the on-disk format upgrade.
Before you upgrade the vSAN on-disk format, run the Pre-Check Upgrade to ensure a smooth upgrade. The pre-check
identifies potential issues that might prevent a successful upgrade, such as failed disks or unhealthy objects.
NOTE
Once you upgrade the on-disk format, you cannot roll back software on the hosts or add certain older hosts to
the cluster.
290
VMware vSAN 8.0
NOTE
If you enable encryption or deduplication and compression on an existing vSAN cluster, the on-disk format is
automatically upgraded to the latest version. This procedure is not required. See Edit vSAN Settings.
1. Navigate to the vSAN cluster.
2. Click the Configure tab.
3. Under vSAN, select Disk Management.
4. (Optional) Click Pre-check Upgrade.
The upgrade pre-check analyzes the cluster to uncover any issues that might prevent a successful upgrade. Some of
the items checked are host status, disk status, network status, and object status. Upgrade issues are displayed in the
Disk pre-check status text box.
5. Click Upgrade.
6. Click Yes on the Upgrade dialog box to perform the upgrade of the on-disk format.
vSAN successfully upgrades the on-disk format. The On-disk Format Version column displays the disk format version of
storage devices in the cluster.
If a failure occurs during the upgrade, you can check the Resyncing Objects page. Wait for all resynchronizations to
complete, and run the upgrade again. You also can check the cluster health using the health service. After you have
resolved any issues raised by the health checks, you can run the upgrade again.
291
VMware vSAN 8.0
• Verify that the hardware and software that you plan on using are certified and listed in the VMware Compatibility Guide
website at http://www.vmware.com/resources/compatibility/search.php.
• Verify that you have enough free space to perform the disk format upgrade. Run the RVC
vsan.whatif_host_failures command to determine that you have enough capacity to complete the upgrade or
perform a component rebuild in case you encounter failure during the upgrade.
• Verify that you have PuTTY or similar SSH client installed for accessing RVC.
For detailed information about downloading the RVC tool and using the RVC commands, see the RVC Command
Reference Guide.
• Verify that your hosts are not in maintenance mode. When upgrading the on-disk format, do not place your hosts in
maintenance mode. When any member host of a vSAN cluster enters maintenance mode, the available resource
capacity in the cluster is reduced because the member host no longer contributes capacity to the cluster. The cluster
upgrade might fail.
• Verify that there are no component rebuilding tasks currently in progress in the vSAN cluster by running the RVC
vsan.resync_dashboard command.
292
VMware vSAN 8.0
This is typically difficult to plan for and hence the guidance was to keep 30 percent of free space in the cluster assuming
that it is unlikely that the largest object in the cluster consumes more than 25 percent of the space and 5 percent of the
space is reserved to make sure cluster does not become full due to policy changes. In vSAN 7.0U1 and later, all objects
are created in a new format which allows the operations space needed by vSAN to perform policy change on an object if
there is 255 GB per host for objects less than 8 TB and 765 GB per host for objects 8 TB or larger.
After a cluster is upgraded to vSAN 7.0 U1 or later from vSAN 7.0 or earlier release, the objects greater than 255 GB
created with the older release must be rewritten in the new format before vSAN can provide the benefit of being able to
perform operations on an object with the new free space requirements. A new object format health alert is displayed after
an upgrade, if there are objects that must be fixed to the new object format and allows the health state to be remediated
by starting a relayout task to fix these objects. The health alert provides information on the number of objects that must
be fixed and the amount of data that will be rewritten. The cluster might experience a drop of about 20 percent in the
performance while the relayout task is in progress. The resync dashboard provides more accurate information about the
amount of time this operation takes to complete.
Using the RVC Upgrade Command Options During vSAN Cluster Upgrade
The vsan.ondisk_upgrade command provides various command options that you can use to control and manage the
vSAN cluster upgrade.
For example, you can allow reduced redundancy to perform the upgrade when you have little free space available. Run
the vsan.ondisk_upgrade --help command to display the list of RVC command options.
Use these command options with the vsan.ondisk_upgrade command.
Options Description
--hosts_and_clusters Use to specify paths to all host systems in the cluster or cluster's compute resources.
--ignore-objects, -i Use to skip vSAN object upgrade. You can also use this command option to eliminate
the object version upgrade. When you use this command option, objects continue to use
the current on-disk format version.
--allow-reduced-redundancy, -a Use to remove the requirement of having a free space equal to one disk group during
disk upgrade. With this option, virtual machines operate in a reduced redundancy
mode during upgrade, which means certain virtual machines might be unable to
tolerate failures temporarily and that inability might cause data loss. vSAN restores full
compliance and redundancy after the upgrade is completed.
--force, -f Use to enable force-proceed and automatically answer all confirmation questions.
--help, -h Use to display the help options.
For information about using the RVC commands, see the RVC Command Reference Guide.
293
VMware vSAN 8.0
294
VMware vSAN 8.0
current vSAN Release Catalog. vSAN also includes the necessary drivers and patch updates for the recommended
release in its system baseline.
vSAN build recommendations ensure that each vSAN cluster remains at the current hardware compatibility status or
better. If hardware in the vSAN cluster is not included on the HCL, vSAN can recommend an upgrade to the latest release,
since it is no worse than the current state.
NOTE
vSphere Lifecycle Manager uses the vSAN health service when performing remediation precheck for hosts
in a vSAN cluster. vSAN health service is not available on hosts running ESXi 6. 0 Update 1 or earlier. When
vSphere Lifecycle Manager upgrades hosts running ESXi 6.0 Update 1 or earlier, the upgrade of the last host
in the vSAN cluster might fail. If remediation failed because of vSAN health issues, you can still complete
the upgrade. Use the vSAN health service to resolve health issues on the host, then take that host out of
maintenance mode to complete the upgrade workflow.
The following examples describe the logic behind vSAN build recommendations.
Example 1
A vSAN cluster is running 6.0 Update 2, and its hardware is included on the 6.0 Update 2 HCL. The HCL lists the hardware
as supported up to release 6.0 Update 3, but not supported for 6.5 and later. vSAN recommends an upgrade to 6.0 Update 3,
including the necessary critical patches for the release.
Example 2
A vSAN cluster is running 6.7 Update 2, and its hardware is included on the 6.7 Update 2 HCL. The hardware is also
supported on the HCL for release 7.0 Update 3. vSAN recommends an upgrade to release 7.0 Update 3.
Example 3
A vSAN cluster is running 6.7 Update 2 and its hardware is not on the HCL for that release. vSAN recommends an upgrade
to 7.0 Update 3, even though the hardware is not on the HCL for 7.0 Update 3. vSAN recommends the upgrade because the
new state is no worse than the current state.
Example 4
A vSAN cluster is running 6.7 Update 2, and its hardware is included on the 6.7 Update 2 HCL. The hardware is also
supported on the HCL for release 7.0 Update 3 and selected baseline preference is patch-only. vSAN recommends an
upgrade to 7.0 Update 3, including the necessary critical patches for the release.
The recommendation engine runs periodically (once each day), or when the following events occur.
• Cluster membership changes. For example, when you add or remove a host.
• The vSAN management service restarts.
• A user logs in to Broadcom Support Portal using a web browser or RVC.
• An update is made to the VMware Compatibility Guide or the vSAN Release Catalog.
The vSAN Build Recommendation health check displays the current build that is recommended for the vSAN cluster. It
also can warn you about any issues with the feature.
System Requirements
vSphere Lifecycle Manager is an extension service in vCenter Server 7.0 and later.
vSAN requires Internet access to update release metadata, to check the VMware Compatibility Guide, and to download
ISO images from Broadcom Support Portal.
vSAN requires valid credentials to download ISO images for upgrades from Broadcom Support Portal. For hosts running
6.0 Update 1 and earlier, you must use RVC to enter the Broadcom Support Portal credentials. For hosts running later
software, you can log in from the ESX Build Recommendation health check.
To enter Broadcom Support Portal credentials from RVC, run the following command: vsan.login_iso_depot -
u<username>-p<password>
295
VMware vSAN 8.0
Intended Audience
This manual is intended for anyone who wants to monitor vSAN operation and performance, or troubleshoot problems
with a vSAN cluster. The information in this manual is written for experienced system administrators who are familiar with
virtual machine technology and virtual datacenter operations. This manual assumes familiarity with VMware vSphere,
including VMware ESXi, vCenter Server, and the vSphere Client.
For more information about vSAN and how to create a vSAN cluster, see the vSAN Planning and Deployment Guide.
For more information about vSAN features and how to configure a vSAN cluster, see Administering VMware vSAN.
What Is vSAN
VMware vSAN is a distributed layer of software that runs natively as a part of the ESXi hypervisor.
vSAN aggregates local or direct-attached capacity devices of a host cluster and creates a single storage pool shared
across all hosts in the vSAN cluster. While supporting VMware features that require shared storage, such as HA, vMotion,
and DRS, vSAN eliminates the need for external shared storage and simplifies storage configuration and virtual machine
provisioning activities.
296
VMware vSAN 8.0
297
1. Navigate to the vSAN cluster.
2. Click the Monitor tab.
3. Under vSAN, click Capacity to view the vSAN capacity information.
VMware vSAN 8.0
have deduplication and compression enabled, you can view the deduplication and compression savings and the
deduplication and compression ratio.
NOTE
vSAN Express Storage Architecture (ESA) does not support deduplication.
Terms Description
298
VMware vSAN 8.0
Category Description
NOTE
PMem only supports VMDK, Non-Volatile Dual In-line Memory Module (NVDIMM), and file system
overhead.
When you enable deduplication and compression, it might take several minutes for capacity updates to be reflected in the
Capacity monitor, as disk space is reclaimed and reallocated. For more information about deduplication and compression,
see "Using Deduplication and Compression" in Administering VMware vSAN.
In vSAN ESA, Usage by Snapshots displays the snapshot usage by the vSAN datastore. You can delete one or more
snapshots and free the used space, thus managing space consumption. To delete a snapshot, right-click the virtual
299
VMware vSAN 8.0
machine > Snapshots > Manage Snapshots. Click Delete to delete a snapshot. Click Delete All Snapshots to delete all
the snapshots. The following are the different usage snapshots available:
Snapshot Description
Container volume snapshots Displays the container volume snapshot usage in the vSAN
datastore.
VMDK snapshots Displays the VMDK snapshot usage in the vSAN datastore.
vSAN file share snapshots Displays the file share snapshot usage in the vSAN datastore.
Current data Displays the usage data that is not included in the snapshot usage
data. You can calculate the current data by subtracting the total
snapshot usage from the total used space.
You can check the history of capacity usage in the vSAN datastore. Click Capacity History, select a time range, and click
Show Results.
The Capacity monitor displays two thresholds represented as vertical markers in the bar chart:
• Operations threshold - Displays the space vSAN requires to perform internal operations in the cluster. If the used
space reaches beyond that threshold, vSAN might not be able to operate properly.
• Host rebuild threshold - Displays the space vSAN requires to tolerate one host failure. If the used space reaches
beyond the host rebuild threshold and the host fails, vSAN might not successfully restore all data from the failed host.
If you enable reserved capacity, the Capacity monitor displays the following:
• Operations reserve - Reserved space in the cluster for internal operations.
• Host rebuild reserve - Reserved space for vSAN to be able to repair in case of single host failure. The Capacity
monitor displays the host rebuild threshold only when the host rebuild reserve is enabled.
If the resynchronization of objects is in progress in a cluster, vSAN displays the capacity used in the capacity chart as
operations usage. In case there is enough free space in the cluster, vSAN might use more space than the operations
threshold for the resyncing operations to complete faster.
Click Configure tab to enable the capacity reserve. You can also click Configure > vSAN > Services to enable the
capacity reserve. For more information on configuring the reserved capacity, see Configure Reserved Capacity for vSAN
Cluster.
In a cluster, if there is more utilization than the host rebuild threshold and the reserved capacity is not enabled, the
capacity chart turns yellow as a warning. If the most consumed host fails, vSAN cannot recover the data. If you enable
the host rebuild reserve, the capacity chart turns yellow at 80% of the host rebuild threshold. If the used space reaches
beyond the operations threshold and the reserved capacity is not enabled, vSAN cannot perform or complete operations
such as rebalance, resync object components due to policy changes, and so on. In that case, the capacity chart turns red
to indicate that the disk usage exceeds the operations threshold. For more information about capacity reserve, see About
Reserved Capacity in vSAN Cluster.
300
VMware vSAN 8.0
301
VMware vSAN 8.0
5. Select the check box of the attached block type or file volumes and click View Performance. You can use the
vSAN cluster performance charts to monitor the workload in your cluster. For more information on the vSAN cluster
performance charts, see View vSAN Cluster Performance.
6. Select the check box on one of the container volumes and click View Container Volume. For more information about
monitoring container volumes, see Monitor Container Volumes in vSAN Cluster.
7. Select the check box on one of the file volumes and click View File Share. For more information about file volume,
see Administering VMware vSAN.
302
VMware vSAN 8.0
These soft reservations prevent the creation of new VMs or powering on VMs if such operations consume the reserved
space. Once the reserved capacity is enabled, vSAN does not prevent powered on VM operations, such as I/O from the
guest operating system or applications from consuming the space even after the threshold limits are reached. After you
enable the reserved capacity, you must monitor the disk space health alerts and capacity usage in the cluster and take
appropriate actions to keep the capacity usage below the threshold limits.
NOTE
The reserved capacity is not supported on a vSAN stretched cluster, cluster with fault domains and nested fault
domains, ROBO cluster, or if the number of hosts in the cluster is less than four.
To enable reserved capacity for the host rebuild, you must first enable the operations reserve. When you enable
operations reserve, vSAN reserves 5% additional capacity in the operations reserve as a buffer to ensure you have time
to react to the capacity fullness before the actual threshold is reached.
vSAN indicates when the capacity usage is high in a cluster. The indications can be in the form of health alerts, capacity
chart turning yellow or red, and so on. Due to the reservation, vSAN might not have enough free space left. This results in
the inability to create VMs or VM snapshots, creating or extending virtual disks, and so on.
NOTE
You cannot enable reserved capacity, if the cluster is at a capacity higher than the specified threshold.
303
VMware vSAN 8.0
304
VMware vSAN 8.0
5. Click to enable or deactivate the operations reserve. On enabling the operations reserve, vSAN ensures that the
cluster has enough space to complete the internal operations.
6. Click to enable or deactivate the host rebuild reserve. On enabling the host rebuild reserve, vSAN provides the
reservation of space to repair data back to compliance following a single host failure. You can enable the host rebuild
reserve only after you enable the operations reserve. After enabling, if you deactivate the operations reserve, the host
rebuild reserve gets automatically deactivated.
7. Select Customize alerts. You can set a customized threshold to receive warning and error alerts. The threshold
percentage is calculated based on the available capacity, which is the difference between the total capacity and the
reserved capacity. If you do not set a customized value, vSAN uses the default thresholds to generate alerts.
8. Click Apply.
When a hardware device, host, or network fails, or if a host is placed into maintenance mode, vSAN initiates
resynchronization in the vSAN cluster. However, vSAN might briefly wait for the failed components to come back online
before initiating resynchronization tasks.
The following events trigger resynchronization in the cluster:
• Editing a virtual machine (VM) storage policy. When you change VM storage policy settings, Virtual SAN might initiate
object recreation and subsequent resynchronization of the objects.
Certain policy changes might cause Virtual SAN to create another version of an object and synchronize it with the
previous version. When the synchronization is complete, the original object is discarded.
Virtual SAN ensures that VMs continue to run and are not interrupted by this process. This process might require
additional temporary capacity.
• Restarting a host after a failure.
• Recovering hosts from a permanent or long-term failure. If a host is unavailable for more than 60 minutes (by default),
Virtual SAN creates copies of data to recover the full policy compliance.
• Evacuating data by using the Full data migration mode before you place a host in maintenance mode.
• Exceeding the utilization threshold of a capacity device. Resynchronization is triggered when capacity device utilization
in the Virtual SAN cluster approaches or exceeds the threshold level of 80 percent.
If a VM is not responding due to latency caused by resynchronization, you can throttle the IOPS used for
resynchronization.
305
VMware vSAN 8.0
Verify that hosts in your vSAN cluster are running ESXi 7.0 or later.
1. Navigate to the vSAN cluster.
2. Select the Monitor tab.
3. Click vSAN.
4. Select Resyncing objects.
5. Track the progress of resynchronization of virtual machine objects.
The Object Repair Time defines the time vSAN waits before repairing a non-compliant object after placing a host in
a failed state or maintenance mode. The default setting is 60 minutes. To change the setting, edit the Object Repair
Timer (Configure > vSAN > Services > Advanced Options).
You can also view the following information about the objects that are resynchronized:
Objects Description
Total resyncing objects Total number of objects to be resynchronized in the vSAN cluster.
Bytes left to resync Data (in bytes) that is remaining before the resynchronization is
complete.
Total resyncing ETA Estimated time left for the resynchronization to complete.
The objects to be resynchronized are categorized as active,
queued, and suspended. The objects that are actively
synchronizing fall in the active category. The objects that are in the
queue for resynchronization are the queued objects. The objects
that were actively synchronizing but are now in the suspended
state falls in the suspended category.
Scheduled resyncing Remaining number of objects to be resynchronized.
You can classify scheduled resyncing into two categories:
scheduled and pending. The scheduled category displays the
objects that are not resyncing because the delay timer has not
expired. Resynchronization of objects starts once the timer
expires. The pending category displays the objects with the
expired delay timer that cannot be resynchronized. This can be
due to insufficient resources in the current cluster or the vSAN
FTT policy set on the cluster not being met.
You can also view the resynchronization objects based on various filters such as Intent and Status. Using Show first,
you can modify the view to display the number of objects.
306
VMware vSAN 8.0
NOTE
To provide enough space for maintenance and reprotection, and to minimize automatic rebalancing events in the
vSAN cluster, consider keeping 30-percent capacity available at all times.
307
VMware vSAN 8.0
minimize the performance impact during the rebalancing activity. Once the rebalancing is complete, you can change the
threshold back to the default 30 percentage.
1. Navigate to the vSAN cluster.
2. Click the Configure tab.
3. Under vSAN, select Services.
4. Click to edit Advanced Options.
308
VMware vSAN 8.0
vSAN displays the performance statistics of all the network I/Os in use. vSAN network diagnostics result appears in the
vCenter Server alerts. The redirection to the related performance charts is available in the vSAN network alerts generated
by the network diagnostics service.
309
VMware vSAN 8.0
vSAN alarms are used for monitoring and troubleshooting performance and networking issues in the vSAN cluster. In
vSAN, these events are known as observations.
VOB ID Description
310
VMware vSAN 8.0
You must have the required privilege level of Alarms.Create Alarm or Alarm.Modify Alarm
1. Navigate to the vSAN cluster.
2. On the Configure tab, select Alarm Definitions and click Add.
3. In the Name and Targets page, enter a name and description for the new alarm.
4. From the Target type drop-down menu, select the type of inventory object that you want this alarm to monitor and
click Next.
Depending on the type of target that you choose to monitor, the summary that follows the Targets, change.
5. In the Alarm Rule page, select a trigger from the drop-down menu.
The combined event triggers are displayed. You can set the rule for a single event only. You must create multiple rules
for multiple events.
6. Click Add Argument to select an argument from the drop-down menu.
a) Select an operator from the drop-down menu.
b) Select an option from the drop-down menu to set the threshold for triggering an alarm.
c) Select severity of the alarm from the drop-down menu. You can set the condition to either Show as Warning or
Show as Critical, but not for both. You must create a separate alarm definition for warning and critical status.
7. Select Send email notifications, to send email notifications when alarms are triggered.
8. In the Email to text box, enter recipient addresses. Use commas to separate multiple addresses.
9. Select Send SNMP traps to send traps when alarms are triggered on a vCenter Server instance.
10. Select Run script to run scripts when alarms are triggered.
11. In the Run this script text box, enter the following script or command:
For this type of command... Enter this...
EXE executable files Full pathname of the command. For example, to run the
cmd.exe command in the C:\tools directory, type:
c:\tools\cmd.exe
BAT batch file Full pathname of the command as an argument to the c:
\windows\system32\cmd.exe command. For example, to run the
cmd.bat command in the C:\tools directory, type:
c:\windows\system32\cmd.exe /c c:\tools\cmd.bat
12. Select an advanced action from the drop-down menu. You can define the advanced actions for virtual machine and
hosts. You can add multiple advanced actions for an alarm.
13. Click Next to set the Reset Rule.
14. Select Reset the alarm to green and click Next to review the alarm definition.
15. Select Enable this alarm to enable the alarm and click Create.
311
VMware vSAN 8.0
You can use Overview to monitor the core health issues of your vSAN cluster. You can also view the following:
• Cluster health score based on the health findings
• View the health score trend for 24 hours
• View the health score trend for a particular period
Ensure that the Historical Health Service is enabled to view details of the Health score trend. Click View Details in the
Health score trend chart to examine the health state of the cluster for a selected time point within 24 hours. Use Custom
to customize the time range as per your requirement.
You can use the vSAN Health findings to diagnose issues, troubleshoot problems, and remediate the problems.
The health findings are classified as follows:
• Unhealthy – Critical or important issue(s) being detected that needs attention.
• Healthy – There are no issues found that needs attention.
• Info – Health findings which may not impact the cluster running state but important for awareness.
• Silenced – Health findings have been silenced without triggering vSAN health alarm by intention.
To troubleshoot an issue, you can sort the findings by root cause to resolve the primary issues initially and then verify if
the impacted issues can also be resolved.
vSAN periodically retests each health finding and updates the results. To run the health findings and update the results
immediately, click the Retest button.
312
VMware vSAN 8.0
If you participate in the Customer Experience Improvement Program (CEIP), you can run health findings and send
the data to VMware for advanced analysis. Click Retest with Online health and then click OK. Online notifications is
enabled by default if the vCenter Server can connect to VMware Analytics Cloud without enrolling CEIP. If you do not
want to participate in CEIP, you can still receive vSAN health notifications for software and hardware issues using Online
notifications.
313
VMware vSAN 8.0
describes the health finding and provides information about how to resolve the issue. You can also view the status
history of the health finding for a given period using History Details tab.
• You can click Silence alert on a health finding, so it does not display any warnings or failures.
• Click Healthy to view health findings that are healthy. Click View Current Result to view the current status of the
health finding. Click View History Details to identify the status history of the health finding for a particular time
period. The status is displayed in green. You can also view the status history of the health finding for a given period
using History Details tab.
314
VMware vSAN 8.0
315
VMware vSAN 8.0
PHM uses these events to generate alarms, which appears in the vSAN Skyline Health. For more information on the
vSAN Skyline Health events, see the VMware knowledge base article at https://knowledge.broadcom.com/external/
article?articleNumber=367770.
When the vSAN performance service is turned on, the cluster summary displays an overview of vSAN performance
statistics, including IOPS, throughput, and latency. You can view detailed performance statistics for the cluster, and for
each host, disk group, and disk in the vSAN cluster. You also can view performance charts for virtual machines and virtual
disks.
316
VMware vSAN 8.0
317
VMware vSAN 8.0
indicating that the network diagnostic mode can be resource-intensive. Ensure that you do not enable it for a longer
duration.
10. Click Apply.
318
VMware vSAN 8.0
• To view PMem performance charts, you must have PMem storage attached to the hosts in the cluster.
1. Navigate to the vSAN cluster.
2. Click the Monitor tab.
3. Under vSAN, select Performance.
4. Select Top
319
VMware vSAN 8.0
based on the I/O latency graph of the cluster, you can select a timestamp and get the top contributors with latency
statistics. You can also select a single contributor and view the latency graph. You have the option to switch
between the combined view and table view.
5. Select VM.
Perform one of the following:
• Select Cluster level metrics to display the aggregated performance metrics for the cluster that you selected.
• Select Show specific VMs to display metrics for all the VMs selected. If you enable Show separate chart by
VMs, vSAN displays separate metrics for all the VMs selected.
Select a time range for your query. vSAN displays performance charts for clients running on the cluster, including
IOPS, throughput, latency, congestions, and outstanding I/Os. The statistics on these charts are aggregated from
the hosts within the cluster. You can also select Real-Time as the time range that displays real-time data that is
automatically refreshed every 30 seconds. The real-time statistics data is retained in the SQL database for seven days
until it gets purged.
6. Select Backend. Select a time range for your query. vSAN displays performance charts for the cluster back-end
operations, including IOPS, throughput, latency, congestions, and outstanding I/Os. The statistics on these charts are
aggregated from the hosts within the cluster.
7. Select File Share and choose a file. Select a time range for your query. Select NFS performance or File system
performance based on the protocol layer performance or file system layer performance that you want to display.
vSAN displays performance charts for vSAN file services, including IOPS, throughput, and latency.
8. Select iSCSI and select an iSCSI target or LUN. Select a time range for your query. vSAN displays performance charts
for iSCSI targets or LUNs, including IOPS, bandwidth, latency, and outstanding I/O.
9. (Optional) Select I/O Insight. For more information on I/O Insight, see Use vSAN I/O Insight.
10. Select vSAN Direct to display the performance data of the vSAN direct disks. Select a time range for your query.
vSAN displays performance charts for vSAN direct, including IOPS, bandwidth, latency, and outstanding I/O.
11. Select PMEM to display the performance data of all VMs placed on the PMem storage. Select a time range for your
query. You can also select Real-time as the time range that displays real time data that is automatically refreshed
every 30 seconds. PMem displays performance charts including IOPS, bandwidth, and latency. For more information
about PMem metrics collection settings, see https://kb.vmware.com/s/article/89100.
12. Click Refresh or Show Results to update the display.
320
VMware vSAN 8.0
At the disk group level, you can view statistics for the disk group. At the disk level, you can view statistics for an individual
storage device.
1. Navigate to the vSAN cluster, and select a host.
2. Click the Monitor tab.
3. Under vSAN, select Performance.
4. Select VM.
• Select Host level metrics to display the aggregated performance metrics for the host that you selected.
• Select Show specific VMs to display metrics for all the VMs selected on the host. If you enable Show separate
chart by VMs, vSAN displays separate metrics for all the VMs selected on the host.
Select a time range for your query. vSAN displays performance charts for clients running on the host, including IOPS,
throughput, latency, congestions, and outstanding I/Os. You can also select Real-Time as the time range that displays
real-time data that is automatically refreshed every 30 seconds. The real-time statistics data is retained in the SQL
database for seven days until it gets purged.
5. In vSAN ESA, select Backend Cache. Select a time range for your query. vSAN displays the performance charts for
the backend cache operations of the host, including the overall backend cache statistics, the overall cache miss by the
different types, cache miss by types for the different transactions, and the catch latency for the different transactions.
6. Select Backend. Select a time range for your query. vSAN displays performance charts for the host back-end
operations, including IOPS, throughput, latency, congestions, outstanding I/Os, and resync I/Os.
7. Perform one of the following:
• Select Disks, and select a disk group. Select a time range for your query. vSAN displays performance charts for
the disk group, including front end (Guest) IOPS, throughput, and latency, as well as overhead IOPS and latency.
It also displays the read-cached hit rate, evictions, write-buffer free percentage, capacity and usage, cache disk
destage rate, congestions, outstanding I/O, outstanding I/O size, delayed I/O percentage, delayed I/O average
latency, internal queue IOPS, internal queue throughput, resync IOPS, resync throughput, and resync latency.
• In vSAN ESA, select Disks, and then select a disk. Select a time range for your query. vSAN displays performance
charts for the disk, including vSAN layer IOPS, throughput, and latency. It also displays the physical or firmware
layer IOPS, throughput, and latency.
8. Select Physical Adapters, and select a NIC. Select a time range for your query. vSAN displays performance charts
for the physical NIC (pNIC), including throughput, packets per second, and packets loss rate.
9. Select Host Network, and select a VMkernel adapter, such as vmk1. Select a time range for your query. vSAN
displays performance charts for all network I/Os processed in the network adapters used by vSAN, including
throughput, packets per second, and packets loss rate.
10. Select iSCSI. Select a time range for your query. vSAN displays performance charts for all the iSCSI services on the
host, including IOPS, bandwidth, latency, and outstanding I/Os.
11. (Optional) Select I/O Insight. For more information on I/O Insight, see Use vSAN I/O Insight.
12. Select vSAN Direct to display the performance data of the vSAN direct disks. Select a time range for your query.
vSAN displays performance charts for vSAN direct, including IOPS, bandwidth, latency, and outstanding I/O.
13. Select PMEM to display the performance data of all VMs placed on the PMem storage. Select a time range for your
query. You can also select Real-time as the time range that displays real time data that is automatically refreshed
321
VMware vSAN 8.0
every 30 seconds. PMem displays the performance charts including IOPS, bandwidth, and latency. For more
information about PMem metrics collection settings, see https://kb.vmware.com/s/article/89100.
14. Click Refresh or Show Results to update the display.
322
VMware vSAN 8.0
vSAN displays performance charts for the VMs in the cluster, including IOPS, throughput, I/O size distribution, I/O latency
distribution, and so on.
You can view metrics for the I/O Insight instance that you created.
323
VMware vSAN 8.0
324
VMware vSAN 8.0
You can rerun an instance, and rename or delete the existing instances.
NOTE
All the ESXi hosts and vCenter Server in the vSAN cluster must be running 7.0 Update 3 or later.
Using the I/O trip analyzer scheduler running 8.0 Update 1 or later, you can set the recurrence for I/O trip analyzer
diagnostic operations. You can either set a one time occurrence or set the recurrence to later. On reaching the recurrence
time, the scheduler automatically collects the results. You can view the results collected within 30 days.
325
VMware vSAN 8.0
NOTE
The I/O trip analyzer supports stretched cluster and multiple VMs (maximum 8 VMs and 64 VMDKs) in one
diagnostic run for a single cluster.
1. Navigate to the vSAN cluster, and select a VM.
2. Click the Monitor tab.
3. Under vSAN, select I/O Trip Analyzer.
4. Click Run New Test.
5. In the Run VM I/O Trip Analyzer Test, select the duration of the test.
6. (Optional) Select Schedule for a future time to schedule the test for a later time. You can either select Start now or
enter a time based on your requirement in the Custom time field. Select the repeat options and click Schedule.
NOTE
You can schedule only a single I/O trip analyzer per cluster. You can schedule another I/O trip analyzer after
deleting the current scheduler. To delete a scheduler, click Schedules > Delete. You can also modify a
schedule that you created. Click Schedules > Edit.
7. Click RUN. The trip analyzer test data is persisted and is available only for 30 days.
NOTE
vSAN does not support I/O trip analyzer for virtual disks in a remote vSAN datastore.
8. Click VIEW RESULT to view the visualized I/O topology.
9. From the Virtual Disks drop-down, select the disk for which you want to view the I/O topology. You can also view the
performance details of the network and the disk groups. Click the edge points of the topology to view the latency
details.
Click the edge points of the topology to view the latency details. If there is a latency issue, click the red icon to focus
on that area.
326
VMware vSAN 8.0
When you click Show Results, vSAN transmits performance data to the vSphere backend analytics server. After
analyzing the data, the vSAN performance diagnostics tool displays a list of issues that might have affected the
benchmark performance for the chosen goal.
You can click to expand each issue to view more details about each issue, such as a list of affected items. You also can
click See More or Ask VMware to display a Knowledge Base article that describes recommendations to address the issue
and achieve your performance goal.
327
VMware vSAN 8.0
vSAN performs an automated upload of the support bundle, and does not allow you to review, obfuscate, or otherwise edit
the contents of your support data prior to it being sent to VMware. vSAN connects to the FTP port 21 or HTTPS port 443
of the target server with the domain name vmware.com, to automatically upload the support bundle.
NOTE
Data collected in the support bundle may be considered sensitive. If your support data contains regulated data,
such as personal, health care, or financial data, you may want to avoid uploading the support bundle.
1. Right-click the vSAN cluster in the vSphere Client.
2. Choose menu vSAN > Upload support bundle...
3. Enter your service request ID and a description of your issue.
4. Click Upload.
Command Description
esxcli vsan network list Verify which VMkernel adapters are used for vSAN
communication.
esxcli vsan storage list List storage disks claimed by vSAN.
esxcli vsan storagepool list List storage pool claimed by vSAN ESA. This command is
applicable only for vSAN ESA cluster.
esxcli vsan cluster get Get vSAN cluster information.
esxcli vsan health Get vSAN cluster health status.
esxcli vsan debug Get vSAN cluster debug information.
The esxcli vsan debug commands can help you debug and troubleshoot the vSAN cluster, especially when vCenter
Server is not available.
Use: esxcli vsan debug {cmd} [cmd options]
328
VMware vSAN 8.0
Debug commands:
Command Description
329
VMware vSAN 8.0
Policy:
stripeWidth: 1
CSN: 1
spbmProfileName: vSAN Default Storage Policy
spbmProfileId: aa6d5a82-1c88-45da-85d3-3d74b91a5bad
forceProvisioning: 0
cacheReservation: 0
proportionalCapacity: [0, 100]
spbmProfileGenerationNumber: 0
hostFailuresToTolerate: 1
Configuration:
RAID_1
Component: 47cbdc58-6928-333f-0c51-020010d5dfa3
Component State: ACTIVE, Address Space(B): 273804165120 (255.00GB),
Disk UUID: 52e95956-42cf-4d30-9cbe-763c616614d5, Disk Name: mpx.vmhba1..
Votes: 1, Capacity Used(B): 373293056 (0.35GB),
Physical Capacity Used(B): 369098752 (0.34GB), Host Name: sc-rdops...
Component: 47cbdc58-eebf-363f-cf2b-020010d5dfa3
Component State: ACTIVE, Address Space(B): 273804165120 (255.00GB),
Disk UUID: 52d11301-1720-9901-eb0a-157d68b3e4fc, Disk Name: mpx.vmh...
Votes: 1, Capacity Used(B): 373293056 (0.35GB),
Physical Capacity Used(B): 369098752 (0.34GB), Host Name: sc-rdops-vm..
Witness: 47cbdc58-21d2-383f-e45a-020010d5dfa3
Component State: ACTIVE, Address Space(B): 0 (0.00GB),
Disk UUID: 52bfd405-160b-96ba-cf42-09da8c2d7023, Disk Name: mpx.vmh...
Votes: 1, Capacity Used(B): 12582912 (0.01GB),
Physical Capacity Used(B): 4194304 (0.00GB), Host Name: sc-rdops-vm...
Type: vmnamespace
Path: /vmfs/volumes/vsan:52134fafd48ad6d6-bf03cb6af0f21b8d/New Virtual Machine
Group UUID: 00000000-0000-0000-0000-000000000000
Directory Name: New Virtual Machine
esxcli vsan debug controller list
Device Name: vmhba1
Device Display Name: LSI Logic/Symbios Logic 53c1030 PCI-X Fusion-MPT Dual Ult..
Used By VSAN: true
PCI ID: 1000/0030/15ad/1976
Driver Name: mptspi
Driver Version: 4.23.01.00-10vmw
Max Supported Queue Depth: 127
esxcli vsan debug limit get
Component Limit Health: green
Max Components: 750
Free Components: 748
Disk Free Space Health: green
Lowest Free Disk Space: 99 %
Used Disk Space: 1807745024 bytes
Used Disk Space (GB): 1.68 GB
Total Disk Space: 107365793792 bytes
Total Disk Space (GB): 99.99 GB
Read Cache Free Reservation Health: green
330
VMware vSAN 8.0
Command Description
^L Redraw screen
Space Update display
h or ? Help; show this text
q Quit
f/F Add or remove fields
o/O Change the order of displayed fields
s Set the delay in seconds between updates
# Set the number of instances to display
E Change the selected entity type
L Change the length of the field
l Limit display to specific node id
. Sort by column, same number twice to change sort order
331
VMware vSAN 8.0
If a host does not meet hardware requirements or experiences other problems, vSAN might fail to configure the host. For
example, insufficient memory on the host might prevent vSAN from being configured.
1. Place the host that causes the failure in Maintenance Mode.
2. Move the host out of the vSAN cluster.
3. Resolve the problem that prevents the host to have vSAN configured.
4. Exit Maintenance Mode.
5. Move the host back into the vSAN cluster.
Host with the vSAN service enabled is not in the vCenter cluster Add the host to the vSAN cluster.
1. Right-click the host, and select Move To.
2. Select the vSAN cluster and click OK.
Host is in a vSAN enabled cluster but does not have vSAN service Verify whether vSAN network is properly configured and enabled
enabled on the host. See vSAN Planning and Deployment.
vSAN network is not configured Configure vSAN network. See vSAN Planning and Deployment.
Host cannot communicate with all other nodes in the vSAN Might be caused by network isolation. See the vSAN Planning and
enabled cluster Deployment documentation.
Found another host participating in the vSAN service which is not Make sure that the vSAN cluster configuration is correct and all
a member of this host's vCenter cluster. vSAN hosts are in the same subnet. See vSAN Planning and
Deployment.
332
VMware vSAN 8.0
Component
Description Recovery Cause
Failure State
Degraded A component is in degraded state vSAN starts rebuilding the affected • Failure of a flash caching device
if vSAN detects a permanent components immediately. • Magnetic or flash capacity
component failure and assumes that device failure
the component is not going to recover • Storage controller failure
to working state.
Absent A component is in absent state if vSAN starts rebuilding absent • Lost network connectivity
vSAN detects a temporary component components if they are not • Failure of a physical network
failure where the component might available within a certain time adapter
recover and restore its working state. interval. By default, vSAN starts • ESXi host failure
rebuilding absent components after
• Unplugged flash caching device
60 minutes.
• Unplugged magnetic disk or
flash capacity device
333
VMware vSAN 8.0
Examine the compliance status and the operational state of a virtual machine object to find how a failure in the cluster
affects the virtual machine.
Compliance Status The compliance status of a virtual machine object indicates whether it meets the
requirements of the assigned VM storage policy.
Operational State The operational state of an object can be healthy or unhealthy. It indicates the type
and number of failures in the cluster.
An object is healthy if an intact replica is available and more than 50 percent of the
object's votes are still available.
An object is unhealthy if an entire replica is not available or less than 50 percent of
the object's votes are unavailable. For example, an object might become unhealthy
if a network failure occurs in the cluster and a host becomes isolated.
To determine the overall influence of a failure on a virtual machine, examine the compliance status and the operational
state. If the operational state remains healthy although the object is noncompliant, the virtual machine can continue using
the vSAN datastore. If the operational state is unhealthy, the virtual machine cannot use the datastore.
334
VMware vSAN 8.0
If a virtual machine uses vSAN storage, its storage accessibility might change according to the type of failure in the vSAN
cluster.
Changes in the accessibility occur when the cluster experiences more failures than the policy for a virtual machine object
tolerates.
As a result from a failure in the vSAN cluster, a virtual machine object might become inaccessible. An object is
inaccessible if a full replica of the object is not available because the failure affects all replicas, or when less than 50
percent of the object's votes are available.
According to the type of object that is inaccessible, virtual machines behave in the following ways:
Virtual machine inaccessibility is not a permanent state. After the underlying issue is resolved, and a full replica and more
than 50 percent of the object's votes are restored, the virtual machine automatically becomes accessible again.
vSAN monitors the performance of each storage device and proactively isolates unhealthy devices.
It detects gradual failure of a storage device and isolates the device before congestion builds up within the affected host
and the entire vSAN cluster.
335
VMware vSAN 8.0
If a disk experiences sustained high latencies or congestion, vSAN considers the device as a dying disk, and evacuates
data from the disk. vSAN handles the dying disk by evacuating or rebuilding data. No user action is required, unless the
cluster lacks resources or has inaccessible objects.
Behavior of vSAN
vSAN responds to the storage device failure in the following ways.
Parameter Behavior
If vSAN detects a disk with a permanent error, it makes a limited number of attempts to revive the disk by unmounting and
mounting it.
When a magnetic disk or flash capacity device fails, vSAN evaluates the accessibility of the objects on the device.
vSAN rebuilds them on another host if space is available and the Primary level of failures to tolerate is set to 1 or more.
336
VMware vSAN 8.0
Behavior of vSAN
vSAN responds to the capacity device failure in the following ways.
Parameter Behavior
Primary level of failures to tolerate If the Primary level of failures to tolerate in the VM storage
policy is equal to or greater than 1, the virtual machine objects are
still accessible from another host in the cluster. If resources are
available, starts an automatic reprotection.
If the Primary level of failures to tolerate is set to 0, a virtual
machine object is inaccessible if one of the object's components
resides on the failed capacity device.
Restore the virtual machine from a backup.
I/O operations on the capacity device stops all running I/O operations for 5-7 seconds until it re-
evaluates whether an object is still available without the failed
component.
If determines that the object is available, all running I/O
operations are resumed.
Rebuilding data examines whether the hosts and the capacity devices can satisfy
the requirements for space and placement rules for the objects
on the failed device or disk group. If such a host with capacity is
available, starts the recovery process immediately because the
components are marked as degraded.
If resources are available, an automatic reprotect will occur.
When a storage pool device fails, vSAN evaluates the accessibility of the objects on the device.
vSAN rebuilds them on another host if space is available and the Primary level of failures to tolerate is set to 1 or more.
Parameter Behavior
Primary level of failures to tolerate If the Primary level of failures to tolerate in the VM storage
policy is equal to or greater than 1, the virtual machine objects are
still accessible from another host in the cluster. If resources are
available, starts an automatic reprotection.
If the Primary level of failures to tolerate is set to 0, a virtual
machine object is inaccessible if one of the object's components
resides on the failed capacity device.
Restore the virtual machine from a backup.
I/O operations on the capacity device stops all running I/O operations for 5-7 seconds until it re-
evaluates whether an object is still available without the failed
component.
If determines that the object is available, all running I/O
operations are resumed.
337
VMware vSAN 8.0
Parameter Behavior
Rebuilding data examines whether the hosts and the capacity devices can satisfy
the requirements for space and placement rules for the objects
on the failed device or disk group. If such a host with capacity is
available, starts the recovery process immediately because the
components are marked as degraded.
If resources are available, an automatic reprotect will occur.
When a flash caching device fails, vSAN evaluates the accessibility of the objects on the disk group that contains the
cache device.
vSAN rebuilds them on another host if possible and the Primary level of failures to tolerate is set to 1 or more.
Behavior of vSAN
vSAN responds to the failure of a flash caching device in the following way:
Parameter Behavior
Primary level of failures to tolerate If the Primary level of failures to tolerate in the VM storage
policy is equal to or greater than 1, the virtual machine objects are
still accessible from another host in the cluster. If resources are
available, starts an automatic reprotection.
If the Primary level of failures to tolerate is set to 0, a virtual
machine object is inaccessible if one of the object's components is
on the failed disk group.
I/O operations on the disk group stops all running I/O operations for 5-7 seconds until it re-
evaluates whether an object is still available without the failed
component.
If determines that the object is available, all running I/O
operations are resumed.
Rebuilding data examines whether the hosts and the capacity devices can satisfy
the requirements for space and placement rules for the objects
on the failed device or disk group. If such a host with capacity is
available, starts the recovery process immediately because the
components are marked as degraded.
If a host stops responding due to failure or reboot of the host, vSAN waits for the host to recover and rebuilds the
components elsewhere in the cluster.
338
VMware vSAN 8.0
Behavior of vSAN
vSAN responds to the host failure in the following way:
Parameter Behavior
Primary level of failures to tolerate If the Primary level of failures to tolerate in the VM storage
policy is equal to or greater than 1, the virtual machine objects are
still accessible from another host in the cluster. If resources are
available, starts an automatic reprotection.
If the Primary level of failures to tolerate is set to 0, a virtual
machine object is inaccessible if the object's components reside
on the failed host.
I/O operations on the host stops all running I/O operations for 5-7 seconds until it re-
evaluates whether an object is still available without the failed
component.
If determines that the object is available, all running I/O
operations are resumed.
Rebuilding data If the host does not rejoin the cluster within 60 minutes, examines
whether some of the other hosts in the cluster can satisfy the
requirements for cache, space and placement rules for the objects
on the inaccessible host. If such a host is available, starts the
recovery process.
If the host rejoins the cluster after 60 minutes and recovery has
started, evaluates whether to continue the recovery or stop it and
resynchronize the original components.
When the connectivity between the hosts in the cluster is lost, vSAN determines the active partition.
vSAN rebuilds the components from the isolated partition on the active partition if the connectivity is not restored.
339
VMware vSAN 8.0
Behavior of vSAN
vSAN responds to a network failure in the following way:
Parameter Behavior
Primary level of failures to tolerate If the Primary level of failures to tolerate in the VM storage
policy is equal to or greater than 1, the virtual machine objects are
still accessible from another host in the cluster. If resources are
available, starts an automatic reprotection.
If the Primary level of failures to tolerate is set to 0, a virtual
machine object is inaccessible if the object's components are on
the isolated hosts.
I/O operations on the isolated hosts stops all running I/O operations for 5-7 seconds until it re-
evaluates whether an object is still available without the failed
component.
If determines that the object is available, all running I/O
operations are resumed.
Rebuilding data If the host rejoins the cluster within 60 minutes, synchronizes the
components on the host.
If the host does not rejoin the cluster within 60 minutes, examines
whether some of the other hosts in the cluster can satisfy the
requirements for cache, space and placement rules for the objects
on the inaccessible host. If such a host is available, starts the
recovery process.
If the host rejoins the cluster after 60 minutes and recovery has
started, evaluates whether to continue the recovery or stop it and
resynchronize the original components.
When a storage controller fails, vSAN evaluates the accessibility of the objects on the disk groups that are attached to the
controller.
vSAN rebuilds them on another host.
Symptoms
If a host contains a single storage controller and multiple disk groups, and all devices in all disk groups are failed, then you
might assume that a failure in the common storage controller is the root cause. Examine the VMkernel log messages to
determine the nature of the fault.
340
VMware vSAN 8.0
Behavior of vSAN
vSAN responds to a storage controller failure in the following way:
Parameter Behavior
Primary level of failures to tolerate If the Primary level of failures to tolerate in the VM storage
policy is equal to or greater than 1, the virtual machine objects are
still accessible from another host in the cluster. If resources are
available, starts an automatic reprotection.
If the Primary level of failures to tolerate is set to 0, a virtual
machine object is inaccessible if the object's components reside
on the disk groups that are connected to the storage controller.
Rebuilding data examines whether the hosts and the capacity devices can satisfy
the requirements for space and placement rules for the objects
on the failed device or disk group. If such a host with capacity is
available, starts the recovery process immediately because the
components are marked as degraded.
A vSAN stretched cluster manages failures that occur due to the loss of a network connection between sites or the
temporary loss of one site.
Network Connection Lost Between Active Sites If the network connection fails between the two active sites, the witness host
and the preferred site continue to service storage operations, and keep data
available. When the network connection returns, the two active sites are
resynchronized.
Secondary Site Fails or Loses Network Connection If the secondary site goes offline or becomes isolated from the preferred
site and the witness host, the witness host and the preferred site continue to
service storage operations, and keep data available. When the secondary
site returns to the cluster, the two active sites are resynchronized.
Preferred Site Fails or Loses Network Connection If the preferred site goes offline or becomes isolated from the secondary site
and the witness host, the secondary site continues storage operations if it
remains connected to the witness host. When the preferred site returns to
the cluster, the two active sites are resynchronized.
Witness Host Fails or Loses Network Connection If the witness host goes offline or becomes isolated from the preferred
site or the secondary site, objects become noncompliant but data remains
available. VMs that are currently running are not affected.
Troubleshooting vSAN
Examine the performance and accessibility of virtual machines to diagnose problems in the vSAN cluster.
341
VMware vSAN 8.0
Verify Drivers, Firmware, Storage I/O Controllers Against the VMware Compatibility Guide
Use the vSAN Skyline Health to verify whether your hardware components, drivers, and firmware are compatible with
vSAN.
Using hardware components, drivers, and firmware that are not compatible with vSAN might cause problems in the
operation of the vSAN cluster and the virtual machines running on it.
The hardware compatibility health findings verify your hardware against the VMware Compatibility Guide. For more
information about using the vSAN Skyline Health, see Monitoring vSAN Skyline Health.
Examining Performance in vSAN Cluster
Monitor the performance of virtual machines, hosts, and the vSAN datastore to identify potential storage problems.
Monitor regularly the following performance indicators to identify faults in vSAN storage, for example, by using the
performance charts in the vSphere Client:
• Datastore. Rate of I/O operations on the aggregated datastore.
• Virtual Machine. I/O operations, memory and CPU usage, network throughput and bandwidth.
You can use the vSAN performance service to access detailed performance charts. For information about using the
performance service, see Monitoring vSAN Performance. For more information about using performance data in a vSAN
cluster, see the Troubleshooting Reference Manual.
After you enable vSAN on a cluster, the datastore is not assembled correctly because of a detected network
misconfiguration.
After you enable vSAN on a cluster, on the Summary tab for the cluster the Network Status for vSAN appears as
Misconfiguration detected.
One or more members of the cluster cannot communicate because of either of the following reasons:
• A host in the cluster does not have a VMkernel adapter for vSAN.
• The hosts cannot connect each other in the network.
Join the members of the cluster to the same network. See vSAN Planning and Deployment.
Virtual Machine Appears as Noncompliant, Inaccessible or Orphaned in the vSAN Cluster
The state of a virtual machine that stores data on a vSAN datastore appears as noncompliant, inaccessible, or orphaned
due to the vSAN cluster failures.
A virtual machine on a vSAN datastore is in one of the following states that indicate a fault in the vSAN cluster.
• The virtual machine is non-compliant and the compliance status of some of its object is noncompliant. See Examine
the Compliance of a Virtual Machine in vSAN.
• The virtual machine object is inaccessible or orphaned. See Examine the Failure State of a Component.
If an object replica is still available on another host, vSAN forwards the I/O operations of the virtual machine to the replica.
If the object of the virtual machine can no longer satisfy the requirement of the assigned VM storage policy, vSAN
considers it noncompliant. For example, a host might temporarily lose connectivity. See Object States That Indicate
Problems in vSAN.
If vSAN cannot locate a full replica or more than 50 percent of the votes for the object, the virtual machine becomes
inaccessible. If a vSAN detects that the .vmx file is not accessible because the VM Home Namespace is corrupted, the
virtual machine becomes orphaned. See Accessibility of Virtual Machines Upon a Failure in vSAN.
342
VMware vSAN 8.0
If the cluster contains enough resources, vSAN automatically recovers the corrupted objects if the failure is permanent.
If the cluster does not have enough resources to rebuild the corrupted objects, extend the space in the cluster. See
Administering VMware vSAN.
Attempt to Create a Virtual Machine on vSAN Fails
When you try to deploy a virtual machine in a vSAN cluster, the operation fails with an error that the virtual machine files
cannot be created.
The operation for creating a virtual machine fails with an error status: Cannot complete file creation
operation.
The deployment of a virtual machine on vSAN might fail for several reasons.
• vSAN cannot allocate space for the virtual machine storage policies and virtual machine objects. Such a failure
might occur if the datastore does not have enough usable capacity, for example, if a physical disk is temporarily
disconnected from the host.
• The virtual machine has very large virtual disks and the hosts in the cluster cannot provide storage for them based on
the placement rules in the VM storage policy
For example, if the Primary level of failures to tolerate in the VM storage policy is set to 1, vSAN must store two
replicas of a virtual disk in the cluster, each replica on a different host. The datastore might have this space after
aggregating the free space on all hosts in the cluster. However, no two hosts can be available in the cluster, each
providing enough space to store a separate replica of the virtual disk.
vSAN does not move components between hosts or disks groups to free space for a new replica, even though the
cluster might contain enough space for provisioning the new virtual machine.
Verify the state of the capacity devices in the cluster.
a) Navigate to the cluster.
b) On the Monitor tab, click vSAN and select Physical Disks.
c) Examine the capacity and health status of the devices on the hosts in the cluster.
vSAN Stretched Cluster Configuration Error When Adding a Host
Before adding new hosts to a vSAN stretched cluster, all current hosts must be connected. If a current host is
disconnected, the configuration of the new host is incomplete.
After you add a host to a vSAN stretched cluster in which some hosts are disconnected, on the Summary tab for the
cluster the Configuration Status for vSAN appears as Unicast agent unset on host.
When a new host joins a stretched cluster, vSAN must update the configuration on all hosts in the cluster. If one or more
hosts are disconnected from the vCenter Server, the update fails. The new host successfully joins the cluster, but its
configuration is incomplete.
Verify that all hosts are connected to vCenter Server, and click the link provided in the Configuration Status message to
update the configuration of the new host.
If you cannot rejoin the disconnected host, remove the disconnected host from the cluster, and click the link provided in
the Configuration Status message to update the configuration of the new host.
vSAN Stretched Cluster Configuration Error When Using RVC to Add a Host
If you use the RVC tool to add a host to a vSAN stretched cluster, the configuration of the new host is incomplete.
After you use the RVC tool to add a host to a vSAN stretched cluster, on the Summary tab for the cluster the Configuration
Status for vSAN appears as Unicast agent unset on host.
343
VMware vSAN 8.0
When a new host joins a stretched cluster, vSAN must update the configuration on all hosts in the cluster. If you use the
RVC tool to add the host, the update does not occur. The new host successfully joins the cluster, but its configuration is
incomplete.
Verify that all hosts are connected to vCenter Server, and click the link provided in the Configuration Status message to
update the configuration of the new host.
Cannot Add or Remove the Witness Host in vSAN Stretched Cluster
Before adding or removing the witness host in a vSAN stretched cluster, all current hosts must be connected. If a current
host is disconnected, you cannot add or remove the witness host.
When you add or remove a witness host in a vSAN stretched cluster in which some hosts are disconnected, the operation
fails with an error status: The operation is not allowed in the current state. Not all hosts in
the cluster are connected to Virtual Center.
When the witness host joins or leaves a stretched cluster, vSAN must update the configuration on all hosts in the cluster.
If one or more hosts are disconnected from the vCenter Server, the witness host cannot be added or removed.
Verify all hosts are connected to vCenter Server, and retry the operation. If you cannot rejoin the disconnected host,
remove the disconnected host from the cluster, and then you can add or remove the witness host.
Disk Group Becomes Locked in vSAN Cluster
In an encrypted vSAN cluster, when communication between a host and the KMS is lost, the disk group can become
locked if the host reboots.
vSAN locks a host's disk groups when the host reboots and it cannot get the KEK from the KMS. The disks behave as if
they are unmounted. Objects on the disks become inaccessible.
You can view a disk group's health status on the Disk Management page in the vSphere Client. An Encryption health
finding warning notifies you that a disk is locked.
Hosts in an encrypted vSAN cluster do not store the KEK on disk. If a host reboots and cannot get the KEK from the KMS,
vSAN locks the host's disk groups.
To exit the locked state, you must restore communication with the KMS and reestablish the trust relationship.
You must replace a flash caching device if you detect a failure or when there is a disk group upgrade.
• Verify that the storage controllers on the hosts are configured in passthrough mode and support the hot-plug feature.
344
VMware vSAN 8.0
If the storage controllers are configured in RAID 0 mode, see the vendor documentation for information about adding
and removing devices.
• If you upgrade the flash caching device, verify the following requirements:
– If you upgrade the flash caching device, verify that the cluster contains enough space to migrate the data from the
disk group that is associated with the flash device.
– Place the host in maintenance mode.
Removing the cache device removes the entire disk group from the vSAN cluster. When you replace a flash caching
device, the virtual machines on the disk group become inaccessible and the components on the group are marked as
degraded. See A Flash Caching Device Is Not Accessible in a vSAN Cluster.
1. Navigate to the vSAN cluster.
2. On the Configure tab, click Disk Management under vSAN.
3. Select the entire disk group that contains the flash caching device that you want to remove. vSAN does not allow you
to remove the cache disk. To remove the cache disk, you must remove the entire disk group.
4.
vSAN removes the flash caching device along with the entire disk group from the cluster.
1. Add a new device to the host.
The host automatically detects the device.
2. If the host is unable to detect the device, perform a device rescan.
For more information on creating a disk group, claiming storage devices, or adding devices to the disk group in the vSAN
Cluster, see Device Management in a vSAN Cluster.
Replace a Capacity Device in vSAN Cluster
You must replace a flash capacity device or a magnetic disk if you detect a failure or when you upgrade it.
• Verify that the storage controllers on the hosts are configured in passthrough mode and support the hot-plug feature.
If the storage controllers are configured in RAID 0 mode, see the vendor documentation for information about adding
and removing devices.
• If you upgrade the capacity device, verify that the cluster contains enough space to migrate the data from the capacity
device.
Before you physically remove the device from the host, you must manually delete the device from vSAN. When you
unplug a capacity device without removing it from the vSAN cluster, the components on the disk are marked as absent. If
the capacity device fails, the components on the disk are marked as degraded. When the number of failures of the object
replica with the affected components exceeds the FTT value, the virtual machines on the disk become inaccessible. See
Capacity Device Not Accessible in vSAN Cluster.
345
VMware vSAN 8.0
NOTE
If your vSAN cluster uses deduplication and compression, you must remove the entire disk group from the
cluster before you replace the device.
You can also watch the video about how to replace a failed capacity device in vSAN.
1. Navigate to the vSAN cluster.
2. On the Configure tab, click Disk Management under vSAN.
3. Select the flash capacity device or magnetic disk, and click Remove Disk.
NOTE
You cannot remove a capacity device from the cluster with enabled deduplication and compression.
You must remove the entire disk group. If you want to remove a disk group from a vSAN cluster with
deduplication and compression enabled, see "Adding or Removing Disks with Deduplication and
Compression Enabled" in Administering VMware vSAN.
4. In the Remove Disk dialog box, select Full data migration to transfer all the data available on the host to other hosts
in the cluster.
5. Click Go To Pre-Check to find the impact on the cluster if the object is removed or placed in maintenance mode.
6. Click Remove to remove the capacity device.
1. Add a new device to the host.
The host automatically detects the device.
2. If the host is unable to detect the device, perform a device rescan.
Replace a Storage Pool Device in vSAN ESA Cluster
The storage pool represents the amount of capacity provided by the host to the vSAN datastore.
• Verify that the storage controllers on the hosts are configured in passthrough mode and support the hot-plug feature.
If the storage controllers are configured in RAID 0 mode, see the vendor documentation for information about adding
and removing devices.
• If you upgrade the storage pool device, verify that the cluster contains enough space to migrate the data from the
storage pool device.
Each host's storage devices claimed by vSAN form a storage pool. All storage devices claimed by vSAN contribute to
capacity and performance.
1. Navigate to the vSAN cluster.
2. On the Configure tab, click Disk Management under vSAN.
3. Select the storage pool device, and click Remove Disk.
4. In the Remove Disk dialog box, select Full data migration to transfer all the data available on the host to other hosts
in the cluster.
5. Click Go To Pre-Check to find the impact on the cluster if the object is removed or placed in maintenance mode.
6. Click Remove to remove the storage pool device.
1. Add a new device to the host.
The host automatically detects the device.
2. If the host is unable to detect the device, perform a device rescan.
3. Claim a disk using the vSAN cluster > Configure > vSAN > Disk Management.
346
VMware vSAN 8.0
If you detect a failed storage device or if you upgrade a device, you can manually remove it from a host by using an
ESXCLI command.
Verify that the storage controllers on the hosts are configured in passthrough mode and support the hot-plug feature.
If the storage controllers are configured in RAID 0 mode, see the vendor documentation for information about adding and
removing devices.
If you remove a flash caching device, vSAN deletes the disk group that is associated with the flash device and all its
member devices.
1. Open an SSH connection to the ESXi host.
2. To identify the device ID of the failed device, run this command and learn the device ID from the output.
esxcli vsan storage list
The following are the commands available for managing vSAN ESA cluster:
Command Description
esxcli vsan storagepool add Add physical disk for vSAN usage.
esxcli vsan storagepool list List vSAN storage pool configuration.
esxcli vsan storagepool mount Mount vSAN disk from storage pool.
esxcli vsan storagepool rebuild Rebuild vSAN storage pool disks.
esxcli vsan storagepool remove Remove physical disk from storage pool. Requires one --disk or --
uuid param.
esxcli vsan storagepool unmount Unmount vSAN disk from storage pool.
347
VMware vSAN 8.0
Shut Down the vSAN Cluster Using the Shutdown Cluster Wizard
Use the Shutdown cluster wizard to gracefully shut down the vSAN cluster for maintenance or troubleshooting.
The Shutdown Cluster Wizard is available with vSAN 7.0 Update 3 and later releases.
NOTE
If you have a vSphere with Tanzu environment, you must follow the specified order when shutting down or
starting up the components. For more information, see "Shutdown and Startup of VMware Cloud Foundation" in
the VMware Cloud Foundation Operations Guide.
1. Prepare the vSAN cluster for shutdown.
a) Check the vSAN Skyline Health to confirm that the cluster is healthy.
b) Power off all virtual machines (VMs) stored in the vSAN cluster, except for vCenter Server VMs, vCLS VMs and
file service VMs. If vCenter Server is hosted on the vSAN cluster, do not power off the vCenter Server VM or VM
service VMs (such as DNS, Active Directory) used by vCenter Server.
c) If this is an HCI Mesh server cluster, power off all client VMs stored on the cluster. If the client cluster's vCenter
Server VM is stored on this cluster, either migrate or power off the VM. Once this server cluster is shutdown, its
shared datastore is inaccessible to clients.
d) Verify that all resynchronization tasks are complete.
Click the Monitor tab and select vSAN > Resyncing Objects.
NOTE
If any member hosts are in lockdown mode, add the host's root account to the security profile Exception User
list. For more information, see Lockdown Mode in vSphere Security.
2. Right-click the vSAN cluster in the vSphere Client, and select menu Shutdown cluster.
You also can click Shutdown Cluster on the vSAN Services page.
348
VMware vSAN 8.0
3. On the Shutdown cluster wizard, verify that the Shutdown pre-checks are green checks. Resolve any issues that are
red exclamations. Click Next.
If vCenter Server appliance is deployed on the vSAN cluster, the Shutdown cluster wizard displays the vCenter Server
notice. Note the IP address of the orchestration host, in case you need it during the cluster restart. If vCenter Server
uses service VMs such as DNS or Active Directory, note them as exceptional VMs in the Shutdown cluster wizard.
4. Enter a reason for performing the shutdown, and click Shutdown.
The vSAN Services page changes to display information about the shutdown process.
5. Monitor the shutdown process.
vSAN performs the steps to shutdown the cluster, powers off the system VMs, and powers off the hosts.
Restart the vSAN cluster. See Restart the vSAN Cluster.
Restart the vSAN Cluster
You can restart a vSAN cluster that is shut down for maintenance or troubleshooting.
349
VMware vSAN 8.0
f) If vCenter Server is hosted on the vSAN cluster, power off the vCenter Server VM.
Make a note of the host that runs the vCenter Server VM. It is the host where you must restart the vCenter Server
VM.
g) Deactivate cluster member updates from vCenter Server by running the following command on the ESXi hosts in
the cluster. Ensure that you run the following command on all the hosts.
esxcfg-advcfg -s 1 /VSAN/IgnoreClusterMemberListUpdates
h) Log in to any host in the cluster other than the witness host.
i) Run the following command only on that host. If you run the command on multiple hosts concurrently, it may cause
a race condition causing unexpected results.
python /usr/lib/vmware/vsan/bin/reboot_helper.py prepare
350
VMware vSAN 8.0
If any hosts fail to restart, you must manually recover the hosts or move the bad hosts out of the vSAN cluster.
b) When all the hosts are back after powering on, exit all hosts from maintenance mode. If the vCenter Server is
powered off, use the following command on the ESXi hosts to exit maintenance mode.
esxcli system maintenanceMode set -e false
Following are the commands to remove and add the host to a vSAN cluster:
#esxcli vsan cluster unicastagent remove -a <IP Address> -t node -u <NodeUuid>
#esxcli vsan cluster unicastagent add -t node -u <NodeUuid> -U true -a <IP Address> -p 12321
351
VMware vSAN 8.0
352