On Thursday, March 27, 2014 5:19:26 PM UTC+1, Nikolaus Rath wrote: > > > PA Nilsson <[email protected] <javascript:>> writes: > > A follow up question on this. > > > > When mounting a file system using an s3c backend, running 'df' will > > report that the filsystem has 1TB size. There is no such information > > coming from the backend, but is is easily made available. > > > > Can information on this somehow be propagated? > > As far as I know, there is no such information from most > backends. Google Storage, Amazon S3, OpenStack et al all have > effectively unlimited storage. I believe only the local backend could > effectively report a capacity. Which backend do you have in mind? > > But even if we got a number from the backend, it's not clear what we > should report to df. How do we take into account compression and > deduplication?
We are using a custom storage with a Amazon S3 like interface, so we will be using a slightly modified s3c backend. However, we will be using quotas in this implementation. But just as you say, it is very hard to account for compression and such, so in my opinion the only thing that can be reported is the actual capacity available. How that capacity is used should really not be of concern from the storage provider, it shall just report the capacity available on the disk. Or am I missing something here? /PA -- You received this message because you are subscribed to the Google Groups "s3ql" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
