Hi Anand, I tested this policy with fio and dstat. It performs overall really well. On my raid1c3 array with two HDDs and one SSD (which is the last device), I'm getting the following results.
Michal, Thank you for verifying. More below...
With direct=0: Run status group 0 (all jobs): READ: bw=3560MiB/s (3733MB/s), 445MiB/s-445MiB/s (467MB/s-467MB/s), io=3129GiB (3360GB), run=900003-900013msec With direct=1: Run status group 0 (all jobs): READ: bw=520MiB/s (545MB/s), 64.9MiB/s-65.0MiB/s (68.1MB/s-68.2MB/s), io=457GiB (490GB), run=900001-900001msec However, I was also running dstat at the same time and I noticed that the read stop sometimes for ~15-20 seconds. For example: ----system---- --dsk/sdb-- --dsk/sdc-- --dsk/sdd-- 20-01 00:37:21| 0 0 : 0 0 : 509M 0 20-01 00:37:22| 0 0 : 0 0 : 517M 0 20-01 00:37:23| 0 0 : 0 0 : 507M 0 20-01 00:37:24| 0 0 : 0 0 : 518M 0 20-01 00:37:25| 0 0 : 0 0 : 22M 0 20-01 00:37:26| 0 0 : 0 0 : 0 0 20-01 00:37:27| 0 0 : 0 0 : 0 0 20-01 00:37:28| 0 0 : 0 0 : 0 0 20-01 00:37:29| 0 0 : 0 0 : 0 0 20-01 00:37:30| 0 0 : 0 0 : 0 0 20-01 00:37:31| 0 0 : 0 0 : 0 0 20-01 00:37:32| 0 0 : 0 0 : 0 0 20-01 00:37:33| 0 0 : 0 0 : 0 0 20-01 00:37:34| 0 0 : 0 0 : 0 0 20-01 00:37:35| 0 0 : 0 0 : 0 0 20-01 00:37:36| 0 0 : 0 0 : 0 0 20-01 00:37:37| 0 0 : 0 0 : 0 0 20-01 00:37:38| 0 0 : 0 0 : 0 0 20-01 00:37:39| 0 0 : 0 0 : 0 0 20-01 00:37:40| 0 0 : 0 0 : 0 0 20-01 00:37:41| 0 0 : 0 0 : 0 0 20-01 00:37:42| 0 0 : 0 0 : 0 0 20-01 00:37:43| 0 0 : 0 0 : 0 0 20-01 00:37:44| 0 0 : 0 0 : 0 0 20-01 00:37:45| 0 0 : 0 0 : 0 0 20-01 00:37:46| 0 0 : 0 0 : 55M 0 20-01 00:37:47| 0 0 : 0 0 : 516M 0 20-01 00:37:48| 0 0 : 0 0 : 515M 0 20-01 00:37:49| 0 0 : 0 0 : 516M 0 20-01 00:37:50| 0 0 : 0 0 : 520M 0 20-01 00:37:51| 0 0 : 0 0 : 520M 0 20-01 00:37:52| 0 0 : 0 0 : 514M 0 Here is the full log: https://susepaste.org/16928336 I never noticed that happening with the PID policy. Is that maybe because of reading the part stats for all CPUs while selecting the mirror?
I ran fio tests again, now with dstat in an another window. I don't notice any such stalls, the read numbers went continuous until fio finished. Could you please check with the below fio command, also could you please share your fio command options. fio \ --filename=/btrfs/largefile \ --directory=/btrfs \ --filesize=50G \ --size=50G \ --bs=64k \ --ioengine=libaio \ --rw=read \ --direct=1 \ --numjobs=1 \ --group_reporting \ --thread \ --name iops-test-job It is system specific? Thanks. Anand
Michal