Hello,
Hoping someone may be experiencing the same issues that we have at one of our customers sites.
We have implemented a 2 node Server 2019 S2D (Azure HCI Stack) solution, the hardware was purchased from Dell and is completely HCI compliant (the nodes we have on the bezels "Storage Ready Nodes", however, I believe its now called Azure HCI Stack).
This is a 2 node ROBO installation, 2 x 25GB adapters for storage networking, back to back (no switches). The VM network and management are in a SET team using 2 x 10GB adapters (going through 10GB switches).
We have implemented these as per the Dell HCI best practice guide;
https://www.dell.com/support/manuals/uk/en/ukbsdt1/microsoft-storage-ready-node-r740xd/robos2dreadynodedg/notes-cautions-and-warnings?guid=guid-5b8de7b7-879f-45a4-88e0-732155904029&lang=en-us
We are seeing no issues in stability at all, the solution is functioning and generally speaking we and the customer are very happy with it.
Of course there is a BUT...
We are finding that certain guest OS's (namely Server 2008 R2) are performing badly on throughput testing (using iPERF). The general performance of the 2008 R2 VMs is great. Booting/Shutting down and running tasks (SQL, Exchange) on them perform brilliantly.
Its when clients (Windows 7) or other guest VMs (on the other node to it) try to copy data its very slow.
I have run lots of iPERF tests and I will list them below;
*** S2D Troubleshooting ***
Node Owner for S2D disk makes no difference to performance, either node can be owner and I get the same results.
Any VM on same node (Doesn't matter what Guest OS) - No Issues (Masking the problem) - Using local BUS.
VM to VM (Servers)
Server 2008 R2 writing to Server 2008 R2 different node - 166MBit/s
Server 2008 R2 writing to Server 2019 different node - 2.65GBit/s
Server 2019 writing to Server 2008 R2 different node - 638MBit/s
Server 2019 writing to Server 2019 different node - 3.23GBit/s
Client to Servers
Windows 7 writing to Server 2008 R2 (any node) - 232MBit/s
Windows 7 writing to Server 2012 R2 (any node) - 175MBit/s
Windows 7 writing to Server 2019 (any node) - 1GBit/s
Servers to Client
Server 2008 R2 (any node) writing to Windows 7 - 820MBit/s
Server 2012 R2 (any node) writing to Windows 7 - 800MBit/s
Server 2019 (any node) writing to Windows 7 - 832MBit/s
All Guest Integration Tools are up to date, I have tried previous versions also.
As you can see its an issues with the older OS's, we are planning on removing these (upgrading or building new) soon. I am wondering if we are the only ones seeing this particular issue? I also have a case open with Dell and they are looking into it.
If it makes any difference this was a migration from VMware to Hyper-V 2019 S2D, VMware Tools removed and the VM disks we converted and VM guests created from scratch. No legacy network cards on any VM.
Disabling VMQ makes no difference, same with any TCP offload, this issue seems tied to the guest OS and its driving me mad!
Any thoughts would be very welcome!