>
> Okay, I tried that. Here are the results:
[snip]
Alright, I've figured it out. It's somewhat counter-intuitive (at least for
me). It looks like the -l (limit) doesn't take into consideration the size
of the directory.
Example:
# /opt/apache2/bin/htcacheclean -v -t -p/www/cache -l300M
Statistics:
size limit 300.0M
total size was 399.9M, total size now 299.9M
total entries was 27344, total entries now 20483
# df -h /www/cache/
Filesystem Size Used Avail Use% Mounted on
tmpfs 1.2G 415M 786M 35% /www/cache
But if I just add up the .header and .data files:
# count=0; for i in `find /www/cache -type f -ls | awk '{print $7}'`; do let
"count=$i + $count";done; echo $count
314566786
Convert bytes into megabytes:
# echo "scale=1;314566786/1048576" | bc
299.9
So I just need to keep playing with my -l (limit) and -d (daemonize
interval) so I can avoid hitting the ceiling.
Thanks for the help Morgan and Dan.
Matt