Jeremy Wadsack wrote:

This is what the Analog cache files are for (see
http://analog.cx/docs/cache.html). You can create a cache file for
each day's results and then write reports by reading in the cache
files. This will certainly improve performance. It may or may not help
memory problems.


First, Thanks to the several replys.

Couple of things have happened, #1 it was ruled that Analog be pulled
from the "production" boxes - Ta-Da no more memory problems. So
each night a cron jobs gzips access_log[yesterday] and ftp's it to a data
crunch'n box.  We've looked at the stats we need and are looking at the
follow DB tables:

   ByHr:   date, hr, num_request    (don't need pages)
   Stat_Code:   date, code, count
   File_Type:   date, type, count
   File_Name:  date, name, count

Then, and I'm just getting started,
select hr, sum(request) from byhr where date like '2003-09-%' group by hr";
or where date between '2003-04-01' and 2003-06-31' second quarter...
or date = '2003-09-01' and hr = 15 :-/
which returns: from my test data


Each unit () represents 1 request.

hour: reqs: ----: -----:

   0  47070        +++++
   1  35709        ++++
   2  27145
   3  1
   .....
   20  94491
   21  87627
   22  83777
   23  73488

Which is exactly what I'm looking for :-) and my DB file is 1224 bytes for a 3 day
test which means I can do a whole year for under 20 Meg which is 1/3 of one days
access_log. We don't need Domain as 100% of the data is "in-house" as is Organisation.
Status is good for 404's 408's and [what'da'L....] 503's [ just saw that one hmmmmm ]
File type = 5 ea. and Directory Report for which data is being used.


Tks
Chuck

+------------------------------------------------------------------------
|  TO UNSUBSCRIBE from this list:
|    http://lists.isite.net/listgate/analog-help/unsubscribe.html
|
|  Digest version: http://lists.isite.net/listgate/analog-help-digest/
|  Usenet version: news://news.gmane.org/gmane.comp.web.analog.general
|  List archives:  http://www.analog.cx/docs/mailing.html#listarchives
+------------------------------------------------------------------------

Reply via email to