Peter Tribble wrote: 
> I'm not worried about the compression effect. Where I see problems is
> backing up million/tens of millions of files in a single 
> dataset. Backing up
> each file is essentially a random read (and this isn't helped by raidz
> which gives you a single disks worth of  random read I/O per vdev). I
> would love to see better ways of backing up huge numbers of files.

It's worth correcting this point... the RAIDZ behavior you mention only
occurs if the read size is not aligned to the dataset's block size.  The
checksum verifier must read the entire stripe to validate the data, but
it does that in parallel across the stripe's vdevs.  The whole block is
then available for delivery to the application.

Although, backing up millions/tens of millions of files in a single
backup dataset is a bad idea anyway.  The metadata searches will kill
you, no matter what backend filesystem is supporting it.

"zfs send" is the faster way of backing up huge numbers of files.  But
you pay the price in restore time.  (But that's the normal tradeoff)

--Joe
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to