Site icon WinCert

Design and Operation Considerations When Using vSAN Fault Domains

<p><span style&equals;"font-weight&colon; 400&semi;">The concept of <&sol;span><a href&equals;"https&colon;&sol;&sol;docs&period;vmware&period;com&sol;en&sol;VMware-vSphere&sol;7&period;0&sol;com&period;vmware&period;vsphere&period;vsan&period;doc&sol;GUID-8491C4B0-6F94-4023-8C7A-FD7B40D0368D&period;html" target&equals;"&lowbar;blank" rel&equals;"noopener"><b><i>fault domains<&sol;i><&sol;b><&sol;a><span style&equals;"font-weight&colon; 400&semi;"> was introduced to protect the cluster from failures at the level of server racks or disk baskets&comma; which are logically grouped into these domains&period; The inclusion of this mechanism leads to the distribution of data to ensure their fault tolerance&comma; not at the level of individual nodes&comma; but at the level of domains&comma; which will allow you to survive the failure of the whole domain &&num;8211&semi; all nodes grouped in it &lpar;for example&comma; a server rack&rpar;&comma; since replicas of objects will necessarily be placed on nodes from different failure domains&period;<&sol;span><&sol;p>&NewLine;<p><img class&equals;"alignnone size-full wp-image-4727" src&equals;"https&colon;&sol;&sol;www&period;wincert&period;net&sol;wp-content&sol;uploads&sol;2022&sol;08&sol;source-4280758&lowbar;1280&period;jpg" alt&equals;"" width&equals;"1280" height&equals;"853" &sol;><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">The minimum fault domain is a disk group&comma; which is a logically related disk unit&period; Each disk group contains two types of media &&num;8211&semi; cache and capacity&period; The system allows using only solid-state disks as cache carriers&comma; and both magnetic and solid-state disks can act as capacity carriers&period; Caching media helps speed up magnetic disks and reduce latency when accessing data&period;<&sol;span><&sol;p>&NewLine;<h5><span style&equals;"font-weight&colon; 400&semi;">How many fault domains are recommended in a vSAN cluster&quest;<&sol;span><&sol;h5>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">The number of fault domains is calculated using the formula&colon; <&sol;span><i><span style&equals;"font-weight&colon; 400&semi;">the<&sol;span><&sol;i> <i><span style&equals;"font-weight&colon; 400&semi;">number of fault domains &equals; 2 &ast; number of failures to tolerate &plus; 1&period;<&sol;span><&sol;i><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">The minimum required by <&sol;span><a href&equals;"https&colon;&sol;&sol;www&period;starwindsoftware&period;com&sol;starwind-virtual-san" target&equals;"&lowbar;blank" rel&equals;"noopener"><b>vSAN<&sol;b><&sol;a> <span style&equals;"font-weight&colon; 400&semi;">is 2 failure domains&comma; each with one or more hosts&comma; although experts advise four because this provides for the potential of a rebuild in the event of failure &lpar;2-3 domains do not allow rebuilding&rpar;&period; Determining the host is a similar method to figuring out how many fault domains there are in order to determine the required number of failures&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">Ideally&comma; there should be the same number of hosts in each fault domain&comma; the hosts should have an identical configuration&comma; and it is recommended to leave the space of one domain empty for the possibility of rebuilding &lpar;for example&comma; 4 domains with 1 failover&rpar;&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">The fault domain system works not only for Mirroring &lpar;RAID-1&rpar; but also for <&sol;span><a href&equals;"https&colon;&sol;&sol;www&period;techtarget&period;com&sol;searchstorage&sol;definition&sol;erasure-coding" target&equals;"&lowbar;blank" rel&equals;"noopener"><b>Erasure Coding<&sol;b><&sol;a><span style&equals;"font-weight&colon; 400&semi;">&period; In this case&comma; each component of the object must be located in different fault domains&comma; and the formula for calculating the number of fault domains changes a minimum of 4 domains for RAID-5 and 6 domains for RAID-6 &lpar;similar to the calculation of the number of hosts for Erasure Coding&rpar;&period;<&sol;span><&sol;p>&NewLine;<h5><span style&equals;"font-weight&colon; 400&semi;">Design and size of fault-tolerant vSAN structures<&sol;span><&sol;h5>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">To cope with host crashes&comma; the PFTT &lpar;Primary level of failures to tolerate&rpar; attribute must be configured in the virtual machine storage policies&period; More calculated failures &&num;8211&semi; more capacity hosts required&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">When connecting cluster hosts to rack servers&comma; you can arrange them in a fault domain to increase fault tolerance&comma; in particular&comma; to withstand switch failures at the top of the rack and power losses&period; Fault domains do this by distributing redundancy components on various servers in separate compute racks&period; Each of them includes at least one host and must meet the hardware requirements&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">Best practice&comma; in this case&comma; is to use at least four fault domains&comma; as for three&comma; some data escape schemes may not be supported and reprotection after a vSAN failure is not guaranteed&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">When you enable fault domains&comma; vSAN applies the active VM storage policy to them&comma; not to individual hosts&period; Note that if a host is not part of a fault domain&comma; vSAN interprets it as a standalone fault domain&period; When increasing capacity and adding hosts&comma; you can use an existing fault domain configuration or define a separate one&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">It is important to correctly balance storage in terms of fault tolerance&period; To achieve this&comma; consider the following&colon;<&sol;span><&sol;p>&NewLine;<ul>&NewLine;<li style&equals;"font-weight&colon; 400&semi;" aria-level&equals;"1"><span style&equals;"font-weight&colon; 400&semi;">Provide a sufficient number of fault domains to satisfy the calculated <&sol;span><a href&equals;"https&colon;&sol;&sol;www&period;yellow-bricks&period;com&sol;2018&sol;03&sol;19&sol;vsan-stretched-cluster-pftt-and-sftt-what-happens-when-a-full-site-fails-and-multiple-hosts-fail&sol;" target&equals;"&lowbar;blank" rel&equals;"noopener"><b>PFTT<&sol;b><&sol;a><span style&equals;"font-weight&colon; 400&semi;"> &lpar;minimum 3&comma; ideally 4 or more&rpar;&semi;<&sol;span><&sol;li>&NewLine;<li style&equals;"font-weight&colon; 400&semi;" aria-level&equals;"1"><span style&equals;"font-weight&colon; 400&semi;">Assign the same number of hosts to each fault domain&semi;<&sol;span><&sol;li>&NewLine;<li style&equals;"font-weight&colon; 400&semi;" aria-level&equals;"1"><span style&equals;"font-weight&colon; 400&semi;">Use hosts with the same configurations&semi;<&sol;span><&sol;li>&NewLine;<li style&equals;"font-weight&colon; 400&semi;" aria-level&equals;"1"><span style&equals;"font-weight&colon; 400&semi;">Allocate one fault domain of free capacity for data recovery after a failure&period;<&sol;span><&sol;li>&NewLine;<&sol;ul>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">Clusters with different host configurations are not as predictable in their performance&period; It decreases&comma; including due to differences in the types of cache devices&period; In addition&comma; they have different maintenance procedures&period; <&sol;span><span style&equals;"font-weight&colon; 400&semi;">In the case of three hosts&comma; there will only be one failure handled&period; In this case&comma; each of the two necessary replicas of the virtual machine data will be located on different hosts&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">For a three-host structure&comma; if one of the hosts goes for maintenance&comma; VMware vSAN will not be able to evacuate data from it&period; Any additional failure is catastrophic in this mode&period; In this situation&comma; it is recommended that you always use the &&num;8220&semi;Ensure accessibility&&num;8221&semi; option when evacuating&period; It ensures that the objects are still available throughout the data migration&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">Two-three-host configurations typically do not comply with the failover policy standard&period; However&comma; there are different vSAN options that can support this config&period;<&sol;span><&sol;p>&NewLine;<p><img class&equals;"alignnone size-full wp-image-4726" src&equals;"https&colon;&sol;&sol;www&period;wincert&period;net&sol;wp-content&sol;uploads&sol;2022&sol;08&sol;board-2440249&lowbar;1280&period;jpg" alt&equals;"" width&equals;"1280" height&equals;"853" &sol;><&sol;p>&NewLine;<h5><span style&equals;"font-weight&colon; 400&semi;">Fault domains for 2-node clusters<&sol;span><&sol;h5>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">A typical vSAN cluster consists of a minimum of three hosts&comma; each contributing to the total capacity&period; For a <&sol;span><a href&equals;"https&colon;&sol;&sol;core&period;vmware&period;com&sol;resource&sol;vsan-2-node-cluster-guide" target&equals;"&lowbar;blank" rel&equals;"noopener"><b>2-node cluster<&sol;b><&sol;a><span style&equals;"font-weight&colon; 400&semi;">&comma; an external host Witness is necessary in the case of VMware vSAN&period; <&sol;span><span style&equals;"font-weight&colon; 400&semi;">VMware vSAN host with 2-host Witness refers to a dispersal where a user sets up a 2-host vSAN cluster on a single server&period; The vSAN 2-node connects to a switch or&comma; in some versions&comma; through a direct connection&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">The vSAN Witness node&comma; which offers a two-node quorum&comma; can be placed on a third server through low bandwidth&sol;high latency lines or alternate infrastructure at the same site&period; <&sol;span><span style&equals;"font-weight&colon; 400&semi;">Every node is set up as a vSAN fault domain&period; Configuration supported&colon; 1 &plus; 1 &plus; 1 &lpar;2 nodes &plus; Witness host vSAN&rpar;&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">Prior to vSAN 7&comma; a dedicated Witness was needed for every 2-node configuration&period; One or more 2-node setups may make use of a shared Witness device example thanks to vSAN 7 Update 1&period; One Witness Appliance can be used by as many as 64 clusters of two nodes&period; This innovation has a significantly simplified design&comma; controls&comma; and operation&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">When VMs are installed on vSAN host 2 clusters with two fault domains&comma; they frequently duplicate data security&comma; with one version of the data on host 1 and a second version of the data on host 2&period; A Witness component is frequently placed on the vSAN Witness hosting or vSAN Witness device&period;<&sol;span><&sol;p>&NewLine;<p><span style&equals;"font-weight&colon; 400&semi;">A complete duplicate of the VM information is still accessible on the backup host in the case of a host or equipment failure&period; The VM continues to be accessible on the vSAN datastore due to the Witness component&&num;8217&semi;s and alternate replica&&num;8217&semi;s ongoing accessibility&period;<&sol;span><&sol;p>&NewLine;<p><em><strong>Disclaimer&colon; This is a sponsored article&period;<&sol;strong><&sol;em><&sol;p>&NewLine;

Exit mobile version