Mandi! B. Smith
In chel di` si favelave...
> I have a ZFS pool as a dedicated Bacula spool.
To be clear: the ZFS pool is only used for the bacula spool? Or 'spool' is
in 'loose' meaning, eg contain the data 'spooled' from other servers that
have to be put on LTO?
I'm fighting also on this, because ZFS is a bad beast and suffer for 'write
amplification'...
Starting from:
https://arstechnica.com/information-technology/2020/05/zfs-101-understanding-zfs-storage-and-performance/
I've done:
1) (i cannot, not tried) destory the pool and rebuild with an higher ashift
(sector size)
2) add a SSD cache disk to the pool (L2ARC); this probably is needed if your
pool is also the data repository.
3) mount a specific ZFS filesystem, with compressione disabled and bigger
blocksize for the bacula spool:
mkdir -p /rpool-backup/bacula/spool
zfs create -o mountpoint=/rpool-backup/bacula/spool -o compression=off -o
recordsize=1M rpool-backup/bacula-spool
chown -R bacula:tape /rpool-backup/bacula
chmod 770 /rpool-backup/bacula /rpool-backup/bacula/spool
4) even if your data is local, enable bacula spool.
5) split your job in many ones, so you can spool different jobs and
'interleave' writing (eg: where one job is writing to the tape, other are
bulding spool)
6) (to be done) snapshot the data repository, mount the snapshot readonly and
do the backup from that.
I hope i was clear and useful...
--
_______________________________________________
Bacula-users mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bacula-users