On 2012-04-26 11:27, Fred Liu wrote:
“zfs 'userused@' properties” and “'zfs userspace' command” are good
enough to gather usage statistics.
Since no one is focusing on enabling default user/group quota now, the
temporarily remedy could be a script which traverse all the users/groups
in the directory tree. Tough it is not so decent.

find /export/home -type f -uid 12345 -exec du -ks '{}' \; | summing-script

I think you could use some prefetch of dirtree traversal,
like a "slocate" database, or roll your own (perl script).
But yes, it does seem like stone age compared to ZFS ;)

Currently, dedup/compression is pool-based right now,

Dedup is pool-wide, compression is dataset-wide, applied to
individual blocks. Even deeper, both settings apply to new
writes after the corresponding dataset's property was set
(i.e. a dataset can have files with mixed compression levels,
as well as both deduped and unique files).

they don’t have
the granularity on file system or user or group level.
There is also a lot of improving space in this aspect.

This particular problem was discussed a number of times back
on OpenSolaris forum. It boiled down to what you actually
want to have accounted and perhaps billed - the raw resources
spent by storage system, or the logical resources accessed
and used by its users?

Say, you provide VMs with 100Gb of disk space, but your dedup
is lucky enough to use 1TB overall for say 100 VMs. You can
bill 100 users for full 100Gb each, but your operations budget
(and further planning, etc.) has only been hit for 1Tb.


zfs-discuss mailing list

Reply via email to