This is what the Analog cache files are for (seeFirst, Thanks to the several replys.
http://analog.cx/docs/cache.html). You can create a cache file for
each day's results and then write reports by reading in the cache
files. This will certainly improve performance. It may or may not help
memory problems.
Couple of things have happened, #1 it was ruled that Analog be pulled from the "production" boxes - Ta-Da no more memory problems. So each night a cron jobs gzips access_log[yesterday] and ftp's it to a data crunch'n box. We've looked at the stats we need and are looking at the follow DB tables:
ByHr: date, hr, num_request (don't need pages) Stat_Code: date, code, count File_Type: date, type, count File_Name: date, name, count
Then, and I'm just getting started,
select hr, sum(request) from byhr where date like '2003-09-%' group by hr";
or where date between '2003-04-01' and 2003-06-31' second quarter...
or date = '2003-09-01' and hr = 15 :-/
which returns: from my test data
Each unit () represents 1 request.
hour: reqs: ----: -----:
0 47070 +++++ 1 35709 ++++ 2 27145 3 1 ..... 20 94491 21 87627 22 83777 23 73488
Which is exactly what I'm looking for :-) and my DB file is 1224 bytes for a 3 day
test which means I can do a whole year for under 20 Meg which is 1/3 of one days
access_log. We don't need Domain as 100% of the data is "in-house" as is Organisation.
Status is good for 404's 408's and [what'da'L....] 503's [ just saw that one hmmmmm ]
File type = 5 ea. and Directory Report for which data is being used.
Tks Chuck
+------------------------------------------------------------------------ | TO UNSUBSCRIBE from this list: | http://lists.isite.net/listgate/analog-help/unsubscribe.html | | Digest version: http://lists.isite.net/listgate/analog-help-digest/ | Usenet version: news://news.gmane.org/gmane.comp.web.analog.general | List archives: http://www.analog.cx/docs/mailing.html#listarchives +------------------------------------------------------------------------
