Dne 13.2.2015 v 13:31 Henrik Krohns napsal(a):
On Fri, Feb 13, 2015 at 11:12:38AM +0100, Jan Hejl wrote:
Dne 13.2.2015 v 09:47 Axb napsal(a):
On 02/13/2015 08:51 AM, Jan Hejl wrote:
Whups. 62M :-D Sorry, my mistake. But still ...
what counts is:

used_memory_human:49.70M


seems amazingly little data.
Why not?
What counts is tokens (and ttl to purge them). It makes absolutely zero
common sense that 10M messages could be processed having <1M tokens in the
DB.

I process only few hundred messages a day:

$ redis-cli info | egrep '(:keys|memory)'
used_memory:26683624
used_memory_human:25.45M
used_memory_rss:31256576
used_memory_peak:27370904
used_memory_peak_human:26.10M
used_memory_lua:109568
db0:keys=266576,expires=266572,avg_ttl=2305894191

Well 10M is spread over 5 same SA machines, each runs redis instance - 2M per machine - about 25% SA results are cached on client side (these are not sent to spamd) - thus it gives 1,5M per machine / SA node (32 cores, 128G memory, LOAD between 8-12, few non-default SA plugins). Each node can do 5M per day without being overloaded.

Redis output:

used_memory:52039312
used_memory_human:49.63M
used_memory_rss:59494400
used_memory_peak:52873144
used_memory_peak_human:50.42M
used_memory_lua:159744
db0:keys=375966,expires=375962,avg_ttl=3894540722

Bayes settings:

bayes_token_ttl                 60d
bayes_seen_ttl                  14d
bayes_auto_expire               1
bayes_auto_learn                0

I don't think that it's neccessary to use a huge token base - sometimes it's about quality, not quantity.

Attachment: smime.p7s
Description: Elektronicky podpis S/MIME

Reply via email to