We're going to build a server with some 1Tb of over 500 million small
files with size from 0,5k to 4k.  I'm wonder if the ufs2 can handle
this kind of system well. From newfs(8) the min block size is 4k. This
is not optimal in our case, a 1k or 0,5k block is more effective IMHO.
I'd be happy if anyone can suggest what does fragment (block/8) in the
ufs2 mean and how this parameter works. I know It's better to read the
full ufs2 specification, but hope that someone here can give a hint.
Please advice with optimizations or tricks.
Thank you very much.

Try zfs on amd64 unless your app doesn't work well with zfs or your
organization doesn't allow current. Current is remarkably stable
taking into account zfs is fairly new and ported from solaris and
running on current. I'm using it on a 8.2 TB nexsan storage and no
crashes during testing and a limited time in production.

Some years ago I used FreeBSD (5.2) as nfs-server (using ufs2) on
approx. 15 partitions ranging from 400 GB to 2 TB in size.  If the
server for some reason had crashed the webservers were unable to
access the nfs-mounted partitions during the period the server did a
snapshot of a partition, in order to perform a background-fsck and
thus our website was down. So ufs2 does not scale well.


When lenity and cruelty play for a kingdom,
the gentlest gamester is the soonest winner.

freebsd-questions@freebsd.org mailing list
To unsubscribe, send any mail to "[EMAIL PROTECTED]"

Reply via email to