Tony Earnshaw wrote:
Tom Allison wrote:

[...]

What about the tokens and the signature from the first instance?

What are the chances that I could do this without doing data integrity damage or suffering other inconsistencies in performance?

If it had been such a good idea, don't you reckon that Jonathan (who after all wrote dspam in the first place) would have "stumbled" on it long ago?

--Tonni


If that was the case then why would I consider the wheel since someone might have stumbled on that one too...

I was working on a few assumptions:

a token is a representation of essentially a regex match in either case, CRM114 or Bayes. Any overlap is purely coincidental.

How you manipulate the tokens, based on history, is dependent upon the method of calculation, markov/chi-square/naive, but they are dependent on the same base history of good/bad messages and good/bad tokens.

So a signature can consist of both naive derived tokens and SPBH derived tokens.
Any learning or correction of that token will be to apply a correction to the historical count (+1/-1) in either case. So the data and it's history remains consistent.

The more variations you can deploy in checking for spam the better the chances that something will get trapped.

The biggest advantage that dspam can provide is a lighter weight naive or chi-square determination, removing some of the more obvious spam via quarantine, followed by the slower CRM114 methodology to further determine what's left over from the bayes determination.

It probably won't work because there just isn't enough data captured about the tokens. But if it was truely a bad idea then why do so many people use multiple filters to capture spam?

Reply via email to