Actually, I think I just figured it out - it's the Microsoft Continuous 
Availability. It slows everything down to a crawl - sounds like that's almost 
by design for some reason.

Thanks!

Respectfully,
 Tyler



Nov 14, 2022, 9:48 AM by tylerphilli...@tutamail.com:

> Hi Vladislav,
>
> If I don't use the Scale-Out File Server, I don't have any issues with iSCSI 
> speeds: if I directly connect the LUN(s) to the individual servers, I get 
> 'full' speed - it just seems the Scale-Out File Server is causing the issue. 
> Very strange of Microsoft (though not totally unexpected!). I can't find too 
> much online about Scale-Out File Server, other than generic setup information.
>
> Thanks!
>
> Respectfully,
>  Tyler
>
>
>
> Nov 14, 2022, 9:20 AM by bub...@hoster-ok.com:
>
>> Hi
>>
>> On Mon, 2022-11-14 at 15:00 +0100, Tyler Phillippe via Users wrote:
>>
>>> Good idea! I setup a RAM disk on both of those systems, let them sync, 
>>> added it to the cluster. 
>>>
>>> One thing I left out (which didn't hit me until yesterday as a possibility) 
>>> is that I have the iSCSI LUN attached to two Windows servers that are 
>>> acting as a Scale-Out File Server. When I copied a file over to the new 
>>> RAMdisk LUN via Scale-Out File Server, I am still getting 10-20MB/s; 
>>> however, when I create a large file to the underlying, shared DRBD on those 
>>> CentOS machines, I am getting about 700+MB/s, which I watched via iostat. 
>>> So, I guess it's the Scale-Out File Server causing the issue. Not sure why 
>>> Microsoft and the Scale-Out File Server is causing the issue - guess 
>>> Microsoft really doesn't like non-Microsoft backing disks
>>>
>>>
>>
>>
>> Not with Microsoft, but with overall iSCSI performance. For the older iSCSI 
>> target - IET - I used to use the following settings:
>> InitialR2T=No 
>> ImmediateData=Yes 
>> MaxRecvDataSegmentLength=65536 
>> MaxXmitDataSegmentLength=65536 
>> MaxBurstLength=262144 
>> FirstBurstLength=131072 
>> MaxOutstandingR2T=2 
>> Wthreads=128 
>> QueuedCommands=32
>>
>> Without that iSCSI LUNs were very slow independently of backing device speed.
>> Probably LIO provides a way to set them up as well.
>>
>> Best,
>> Vladislav
>>
>>
>>> Does anyone have any experience with that, perhaps? Thanks!!
>>>
>>> Respectfully,
>>>  Tyler
>>>
>>>
>>>
>>> Nov 14, 2022, 2:30 AM by ulrich.wi...@rz.uni-regensburg.de:
>>>
>>>> Hi!
>>>>
>>>> If you have planty of RAM you could configure an iSCSI disk using a ram 
>>>> disk and try how much I/O you get from there.
>>>> Maybe you issue is not-su-much DRBD related. However when my local 
>>>> MD-RAID1 resyncs with about 120MB/s (spinning disks), the system also is 
>>>> hardly usable.
>>>>
>>>> Regards,
>>>> Ulrich
>>>>
>>>>>>> Tyler Phillippe via Users <users@clusterlabs.org> schrieb am 13.11.2022 
>>>>>>> um
>>>>>>>
>>>> 19:26 in Nachricht <ngme_x7--...@tutamail.com>:
>>>>
>>>>> Hello all,
>>>>>
>>>>> I have setup a Linux cluster on 2x CentOS 8 Stream machines - it has 
>>>>> resources to manage a dual primary, GFS2 DRBD setup. DRBD and the cluster 
>>>>> have a diskless witness. Everything works fine - I have the dual primary 
>>>>> DRBD 
>>>>> working and it is able to present an iSCSI LUN out to my LAN. However, 
>>>>> the 
>>>>> DRBD write speed is terrible. The backing DRBD disks (HDD) are RAID10 
>>>>> using 
>>>>> mdadm and they (re)sync at around 150MB/s. DRBD verify has been limited 
>>>>> to 
>>>>> 100MB/s, but left untethered, it will get to around 140MB/s. If I write 
>>>>> data 
>>>>> to the iSCSI LUN, I only get about 10-15MB/s. Here's the DRBD 
>>>>> global_common.conf - these are exactly the same on both machines:
>>>>>
>>>>> global {
>>>>> usage-count no;
>>>>> udev-always-use-vnr;
>>>>> }
>>>>>
>>>>> common {
>>>>> handlers {
>>>>> }
>>>>>
>>>>> startup {
>>>>> wfc-timeout 5;
>>>>> degr-wfc-timeout 5;
>>>>> }
>>>>>
>>>>> options {
>>>>> auto-promote yes;
>>>>> quorum 1;
>>>>> on-no-data-accessible suspend-io;
>>>>> on-no-quorum suspend-io;
>>>>> }
>>>>>
>>>>> disk {
>>>>> al-extents 4096;
>>>>> al-updates yes;
>>>>> no-disk-barrier;
>>>>> disk-flushes;
>>>>> on-io-error detach;
>>>>> c-plan-ahead 0;
>>>>> resync-rate 100M;
>>>>> }
>>>>>
>>>>> net {
>>>>> protocol C;
>>>>> allow-two-primaries yes;
>>>>> cram-hmac-alg "sha256";
>>>>> csums-alg "sha256";
>>>>> verify-alg "sha256";
>>>>> shared-secret "secret123";
>>>>> max-buffers 36864;
>>>>> rcvbuf-size 5242880;
>>>>> sndbuf-size 5242880;
>>>>> }
>>>>> }
>>>>>
>>>>> Respectfully,
>>>>> Tyler
>>>>>
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Manage your subscription:
>>>> https://lists.clusterlabs.org/mailman/listinfo/users
>>>>
>>>> ClusterLabs home: https://www.clusterlabs.org/
>>>>
>>>
>>> _______________________________________________
>>> Manage your subscription:
>>> https://lists.clusterlabs.org/mailman/listinfo/users
>>>
>>> ClusterLabs home: >>> https://www.clusterlabs.org/
>>>
>>
>>
>>
>
>

_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/

Reply via email to