An Slog (you wrote ZIL but you meant Slog as ZIL is onpool logging while Slog is logging on a dedicated device) is not a write cache. It's a logging feature when sync write is enabled and only read after a crash on next bootup. CIFS does not use sync per default and for NFS (that wants sync per default) you can and should disable when you use NFS as a pure backup target. Your benchmarks clearly show that sync is not enabled otherwise write performance with a large vdev would be more like 30-50 MB/s instead your 400-700 MB/s.

If you want to enable sync, you should look at Intel Optane as Slog as this is far better than any other Flash based Slog.

ZFS use RAM as read and write cache. The default write cache is 10% of RAM up to 4GB so the first option to improve write (and read) performance is to add more RAM. A fast L2Arc (ex Intel Optane, size 5-max 10x RAM) can help if you cannot increase RAM or if you want a read ahead functionality that you can enable on an L2Arc. Even write performance is improved with a larger read cache as even writes need to read metadata.

Beside that, I would not create a raid Zn vdev from 24 disks. I would prefer 3 vdevs from 8 disks or at least two vdevs from 12 disks as pool iops scale with number of vdevs.


Gea
@napp-it.org

Am 18.06.2018 um 08:50 schrieb priyadarshan:
On 18 Jun 2018, at 08:27, Oliver Weinmann <oliver.weinm...@icloud.com> wrote:

Hi,

we have a HGST4u60 SATA JBOD with 24 x 10TB disks. I just saw that back then 
when we created the pool we only cared about disk space and so we created a 
raidz2 pool with all 24disks in one vdev. I have the impression that this is 
cool for disk space but is really bad for IO since this only provides the IO of 
a single disk. We only use it for backups and cold CIFS data but I have the 
impression that especially running a single VEEAM backup copy job really maxes 
out the IO. In our case the VEEAM backup copy job reads and writes the data 
from the storage. Now I wonder if it makes sense to restructure the Pool. I 
have to admit that I don't have any other system with a lot of disk space so I 
can't simply mirror the snapshots to another system and recreate the pool from 
scratch.

Would adding two ZIL SSDs improve performance?

Any help is much appreciated.

Best Regards,
Oliver
Hi,

I would be interested to know as well.

Sometimes we have same issue: need for large space vs need to optmise for speed 
(read, write, or both). We also are using, at the moment, 10TB disks, although 
never do RAID-Z2 with more than 10 disks.

This page has some testing that was useful to us: 
https://calomel.org/zfs_raid_speed_capacity.html

Section «Spinning platter hard drive raids» has your use case (although 4TB, 
not 10TB):

24x 4TB, 12 striped mirrors,   45.2 TB,  w=696MB/s , rw=144MB/s , r=898MB/s
24x 4TB, raidz (raid5),        86.4 TB,  w=567MB/s , rw=198MB/s , r=1304MB/s
24x 4TB, raidz2 (raid6),       82.0 TB,  w=434MB/s , rw=189MB/s , r=1063MB/s
24x 4TB, raidz3 (raid7),       78.1 TB,  w=405MB/s , rw=180MB/s , r=1117MB/s
24x 4TB, striped raid0,        90.4 TB,  w=692MB/s , rw=260MB/s , r=1377MB/s

Different adapters/disks will change the results, but I do not thing ratio will 
change much.

It would be interesting to see how ZIL would affect that.


Priyadarshan
_______________________________________________
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss

_______________________________________________
OmniOS-discuss mailing list
OmniOS-discuss@lists.omniti.com
http://lists.omniti.com/mailman/listinfo/omnios-discuss

Reply via email to