Hi,

I have a server with a shared repository for files.  I plan to devote
70GB of an 80GB HD (a single data partition) to the files.  The files are
uploaded and placed in the repository via a web application.  Here is what
I'd like to accomplish:

When directory size exceeds 70GB, delete files,
First-In-First-Out, until the repository is pared back to 70GB.

The best
case scenario would be to pare back the files each time a new file is added.
 However, I am hoping to do this without adding web application logic, which
could cause additional latency for the user.  Although it risks possibly exceeding
the size limit, I am thinking of using a bash script scheduled with cron.
 To ensure against exceeding the limit, I'm leaving 10GB of the 80GB as buffer.
 I know this is imperfect but my humble intellect can't think of another approach.


So I'm looking for comments on two things:

1.  How to make a bash script
look at total directory size, then proceed to delete files FIFO until a target
size is reached;

2.  Whether there is a better alternative than putting
this script on cron.

Thanks!

Bob Shepherd
-- 
TriLUG mailing list        : http://www.trilug.org/mailman/listinfo/trilug
TriLUG Organizational FAQ  : http://trilug.org/faq/
TriLUG Member Services FAQ : http://members.trilug.org/services_faq/
TriLUG PGP Keyring         : http://trilug.org/~chrish/trilug.asc

Reply via email to