Who's paying for this bandwith?

N.

On 4/24/13, Neil Bothwick <n...@digimed.co.uk> wrote:
> On Wed, 24 Apr 2013 19:07:05 +0100, Stroller wrote:
>
>> > That only works on small systems. I have systems here where a 'du' on
>> > /home would take hours and produce massive IO wait, because there's so
>> > much data in there.
>>
>> Of course. Excuse me.
>>
>> My original idea was in respect of the previous respondent's desire to
>> offer hard limits of a gigabyte - allocating each user a partition and
>> running `du`, which returns immediately, on it.
>
> I said "by the gigabyte" not "of a gigabyte", a user could have hundreds
> of them.
>
>> I don't understand how a hard limit could be enforced if it's
>> impractical to assess the size of used data.
>
> Because the filesystem keeps track of the usage, just like it does for
> the whole filesystem, which is why "df ." is so much faster than
> "du .". ZFS does this too, it just doesn't have a concept of a soft limit.
>
>
> --
> Neil Bothwick
>
> Please rotate your phone 90 degrees and try again.
>

Reply via email to