Forum Discussion
SMB Bandwidth Limit for VM Live Migration Category.
Hi All,
PART 1:
Need help to understand the Math or how to calculate the VM Live Migration -SMB Bandwidth Limit?
My Environment Details:
- Considering RDMA capable - Dual NICs of 25GBPS
- Network model of disaggregated, with Storage VLAN1 and Storage VLAN 2 for storage traffic. (No Teaming for Storage adapters).
Considering the above Hardware specs what should be the VM Live migration -bandwidth be?
I have seen this article Optimizing Hyper-V Live Migrations on an Hyperconverged Infrastructure | Microsoft Community Hub. However, it says a limit of 750mbps. I understand that this is for 10gbps Dual NICs. (Manual Set values and no using Overrides via Network ATC).
If there is a similar optimized value for 25gbps NICS, what is the optimized value and how is it calculated?
I see the below table in Microsoft Docs but understand that this is for Converged Model and not for disaggregated.
Is there any resources like this for Disaggregated Network Design?
Part 2:
We have tried to do this via Global overrides to maintain a symmetry across the Cluster.
We have used $Globaloverride.SMBVMlivemigrationlimit in GB. (We did Test out 4, 5,5 GB values)
We cannot add in Decimal values like 3.25 or 3.5 GB here. It's taking only whole numbers.
How do we tackle this? Is this Ideal way to doing it or any better procedure exists without the hassle of manually doing it on all Nodes and ATC Them defaulting it to its own values?
And when we do a clean 23H2 Installation Cloud Based, we see the below value:
How is this calculated?
Part 3:
Is validate-dcb a Valid procedure to validate the 23H2 Clusters?
If Yes, are there any guidelines to tweak the tests in DCB validate Module to suit our requirements?
As per the code, can see that the VM Speed is set to 10gig and I have 25gig NIC cards? Any Knowledge articles that can help with this or any update from Microsoft on validating the 23H2 Clusters?
Sorry for making the Post very long but have been fighting demons of questions related to the above topics in my mind. Any help is greatly Appreciated. Thanks in advance.
1 Reply
Deepak_Kathula Network ATC would limit the SMB Live Migration traffic to 50% of the available bandwidth. Me and other find this overly aggressive.
DCB QoS (RDMA) will handle that storage traffic is ALWAYS PRIORITIZED over any other traffic.
I cannot resonate with the 50% default, when the storage traffic is non-converged, as in most of the S2D, Azure Local Clusters.
Please have a read (TLDR) and especially recommendations below. Does this help you?