On Fri, 13 Mar 2009, decoder wrote:

You create one model file once by feeding it a large corpus of ham+spam.

The problem is that feeding does not work with an SVM algorithm. You have to train on the _whole_ set _always_, so feeding mails is unpractical.

That's why you do this process _once_ with a lot of ham and spam. You can repeat this process any time but it isn't necessary to do this permanently.

I assume it learns from full message corpa? And all it cares about is the rules that hit?

Per my earlier suggestion of learning off the logs + corpa to fix FP/FN, could there be an option to learn off generated minimal corpa files, with their structure being just the rules hit per message (msgid + hits on one possibly very long line)? e.g.:

<kggbph.617...@localhost> 
BAYES_99,FORGED_RCVD_HELO,L_SOME_STD_PROBS,RAZOR2_CF_RANGE_51_100,RAZOR2_CF_RANGE_E4_51_100,RAZOR2_CF_RANGE_E8_51_100,RAZOR2_CHECK,RBL_PSBL_01,RCVD_IN_BRBL,RCVD_IN_NJABL_SPAM,SARE_FROM_SPAM_MONEY2,STOX_30,URIBL_BLACK,URIBL_JP_SURBL,URIBL_WS_SURBL

Then an external tool could generate and maintain these files from the SA log and the maintained training corpa, omitting FP/FN from the log data.

This is just intended to include in training the high- and low-scoring (obviously spam/ham) messages, which may not appear in the training corpa if training is mostly exception-based.

--
 John Hardin KA7OHZ                    http://www.impsec.org/~jhardin/
 jhar...@impsec.org    FALaholic #11174     pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
-----------------------------------------------------------------------
  It is not the place of government to make right every tragedy and
  woe that befalls every resident of the nation.
-----------------------------------------------------------------------
 Tomorrow: Albert Einstein's 130th Birthday

Reply via email to