Re: New spam / phishing rule?

2014-11-08 Thread LuKreme
On Nov 7, 2014, at 10:03 AM, Benny Pedersen m...@junc.eu wrote:
 
 What mua clients shows invalid mimetypes ?

Most all of them.

-- 
He'd never asked for an exciting life. What he really liked, what he
sought on every occasion, was boredom. The trouble was that boredom
tended to explode in your face. Just when he thought he'd found it he'd
be suddenly involved in what he supposed other people - thoughtless,
feckless people - would call an adventure. And he'd be forced to visit
many strange lands and meet exotic and colourful people, although not
for very long because usually he'd be running. He'd seen the creation of
the universe, although not from a good seat, and had visited Hell and
the afterlife. He'd been captured, imprisoned, rescued, lost and
marooned. Sometimes it had all happened on the same day.



Re: New spam / phishing rule?

2014-11-08 Thread Reindl Harald


Am 09.11.2014 um 00:51 schrieb LuKreme:

On Nov 7, 2014, at 10:03 AM, Benny Pedersen m...@junc.eu wrote:


What mua clients shows invalid mimetypes ?


Most all of them


thank you for your fortune footer in the name of everybody trying to 
train ham messages for bayes..


what is that garbage worth for?
__

He'd never asked for an exciting life. What he really liked, what he
sought on every occasion, was boredom. The trouble was that boredom
tended to explode in your face. Just when he thought he'd found it he'd
be suddenly involved in what he supposed other people - thoughtless,
feckless people - would call an adventure. And he'd be forced to visit
many strange lands and meet exotic and colourful people, although not
for very long because usually he'd be running. He'd seen the creation of
the universe, although not from a good seat, and had visited Hell and
the afterlife. He'd been captured, imprisoned, rescued, lost and
marooned. Sometimes it had all happened on the same day.



signature.asc
Description: OpenPGP digital signature


Re: New spam / phishing rule?

2014-11-08 Thread Dave Pooser
On 11/8/14, 5:57 PM, Reindl Harald h.rei...@thelounge.net wrote:

what is that garbage worth for?

It's from a book by Terry Pratchett. Are we really so hard up for things
to talk about that we're going to have a .sig flamewar now?
-- 
Dave Pooser
Cat-Herder-in-Chief, Pooserville.com
Programming: The profession of progressing from WTF? to Oh, duh.




Re: New spam / phishing rule?

2014-11-08 Thread Reindl Harald


Am 09.11.2014 um 01:48 schrieb Dave Pooser:

On 11/8/14, 5:57 PM, Reindl Harald h.rei...@thelounge.net wrote:


what is that garbage worth for?


It's from a book by Terry Pratchett. Are we really so hard up for things
to talk about that we're going to have a .sig flamewar now?


it's not a matter of hard
it's a matter of sending 1 line followed by 10 or more garbage



signature.asc
Description: OpenPGP digital signature


Re: New spam / phishing rule?

2014-11-08 Thread John Hardin

On Sun, 9 Nov 2014, Reindl Harald wrote:


Am 09.11.2014 um 01:48 schrieb Dave Pooser:

 On 11/8/14, 5:57 PM, Reindl Harald h.rei...@thelounge.net wrote:

  what is that garbage worth for?

 It's from a book by Terry Pratchett. Are we really so hard up for things
 to talk about that we're going to have a .sig flamewar now?


it's not a matter of hard
it's a matter of sending 1 line followed by 10 or more garbage


Yep. .sig flamewar. Sigh.

--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Activist: Someone who gets involved.
  Unregistered Lobbyist: Someone who gets involved with something
the MSM doesn't approve of.   -- WizardPC
---
 3 days until Veterans Day


Re: New spam / phishing rule?

2014-11-08 Thread LuKreme
On Nov 8, 2014, at 5:54 PM, Reindl Harald h.rei...@thelounge.net wrote:
 Am 09.11.2014 um 01:48 schrieb Dave Pooser:
 On 11/8/14, 5:57 PM, Reindl Harald h.rei...@thelounge.net wrote:
 
 what is that garbage worth for?
 
 It's from a book by Terry Pratchett. Are we really so hard up for things to 
 talk about that we're going to have a .sig flamewar now?
 
 it's not a matter of hard
 it's a matter of sending 1 line followed by 10 or more garbage

Feel free to bin/killfile/spam-tag my my posts. Makes less than no difference 
to me. That would be far more effective than complaining about it.

-- 
'It's a lovely morning, lads,' he said. 'I feel like a million dollars.
Don't you?' There was a murmur of reluctant agreement. 'Good,' said
Cohen. 'Let's go and get some.' --Interesting Times



Re: New spam / phishing rule?

2014-11-08 Thread Benny Pedersen

On November 9, 2014 2:12:16 AM John Hardin jhar...@impsec.org wrote:


Yep. .sig flamewar. Sigh.


Thats why i use no sig at all, please dont copy me :)


Re: New spam / phishing rule?

2014-11-07 Thread Axb

On 11/07/2014 05:41 PM, David F. Skoll wrote:

Hi,

I've seen a couple of hundred phishing emails come in that all had an
attachment of type application/html which is (of course) bogus.

I've put in a rule to block these and will see how it goes.
I've put an example up at http://pastebin.com/M3dRp4dD
with only slight editing to hide the actual recipient's name.


commiting a mimeheader rule to my sandbox...



Re: New spam / phishing rule?

2014-11-07 Thread Benny Pedersen

On November 7, 2014 5:41:30 PM David F. Skoll d...@roaringpenguin.com wrote:


I've seen a couple of hundred phishing emails come in that all had an
attachment of type application/html which is (of course) bogus.


What mua clients shows invalid mimetypes ?


MUAs and invalid MIME type handling (was Re: New spam / phishing rule?)

2014-11-07 Thread David F. Skoll
On Fri, 07 Nov 2014 18:03:32 +0100
Benny Pedersen m...@junc.eu wrote:

 What mua clients shows invalid mimetypes ?

Microsoft, thank you... if the attachment name ends in .htm or .html it
is treated as HTML regardless of MIME type.

Actually, most MUAs do this.  There are an unbelievable number of MIME
generators that attach everything (PDFs, spreadsheets, whatever) as
application/octet-stream so MUAs are forced to guess the real MIME type
based on the filename or based on sniffing the content. :(

The current state of email sucks, in case nobody's noticed.

Regards,

David.



Re: MUAs and invalid MIME type handling (was Re: New spam / phishing rule?)

2014-11-07 Thread Benny Pedersen

On November 7, 2014 6:06:40 PM David F. Skoll d...@roaringpenguin.com wrote:


 What mua clients shows invalid mimetypes ?
Microsoft, thank you... if the attachment name ends in .htm or .html it
is treated as HTML regardless of MIME type.


Microsoft could fix this in a monthly bugfix update for dangerous software 
fix :)



Actually, most MUAs do this.  There are an unbelievable number of MIME
generators that attach everything (PDFs, spreadsheets, whatever) as
application/octet-stream so MUAs are forced to guess the real MIME type
based on the filename or based on sniffing the content. :(


I think i will submit this as my first mime signature to clamav and 
hopefull get it signed with clamav team as a good signature, so clients 
stop using badly writed software



The current state of email sucks, in case nobody's noticed.


I have, currently i think there is a bug in libspf2 1.2.10, used here in 
opendmarc 1.3.0, problem i see is that pypolicyd-spf does not agre with 
same results as libspf2, hmm


And its weekend here :)


Re: New spam rule for specific content

2013-08-12 Thread Kris Deugau
Amir 'CG' Caspi wrote:
  My main feeling is that if anyone is
 sending HTML email with LOTS of stuff commented out, that email is
 almost certainly spam.  Ham HTML email would probably be done with more
 care.

*snigger*  Take a look at the raw source from a message sent with
Outlook (especially one with stationery) and say that again...

I've had to heavily alter or outright discard a number of otherwise
useful rules along the lines discussed in this thread due to Outlook FPs.

-kgd


Re: New spam rule for specific content

2013-08-12 Thread John Hardin

On Mon, 12 Aug 2013, Kris Deugau wrote:


Amir 'CG' Caspi wrote:

 My main feeling is that if anyone is
sending HTML email with LOTS of stuff commented out, that email is
almost certainly spam.  Ham HTML email would probably be done with more
care.


*snigger*  Take a look at the raw source from a message sent with
Outlook (especially one with stationery) and say that again...

I've had to heavily alter or outright discard a number of otherwise
useful rules along the lines discussed in this thread due to Outlook FPs.


This was my worry, too.

In a word: Microsoft

--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Liberals love sex ed because it teaches kids to be safe around their
  sex organs. Conservatives love gun education because it teaches kids
  to be safe around guns. However, both believe that the other's
  education goals lead to dangers too terrible to contemplate.
---
 3 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-11 Thread Amir 'CG' Caspi

At 1:41 PM -0600 08/10/2013, Amir 'CG' Caspi wrote:
(The HTML comment gibberish rule would be a big step here, since 
that's one of the few things that would distinguish this from ham... 
unlikely that a real person would embed tens of KB of comment 
gibberish.)


OK, I'm trying to test an HTML comment gibberish rule and having some 
problems.  I'm using the following test spam, the same I showed 
before:

http://pastebin.com/VCtvzjzV

I'm testing the following rule:

# HTML comment gibberish
rawbody HTML_COMMENT_GIBBERISH  /!--\s*(?:[\w'?.:;-]+\s+){100,}\s*--/im
tflags HTML_COMMENT_GIBBERISH   multiple
describe HTML_COMMENT_GIBBERISH lots of spammy text in HTML comment
score HTML_COMMENT_GIBBERISH0.001

Now, when I run this test spam through SA, I do get a hit, but only a 
single hit... the rule is popping for the final HTML comment (the one 
beginning with Simpsons).  However, there are two other HTML 
comments in this email, prior to the one that hit... for some reason, 
they are not hitting, even though I've set tflags=multiple.  (I was 
considering having a meta rule that scored extra for multiple 
comments.)


My regex is valid and appropriate for those comments... I tested it 
at regexpal.com, which shows that all three comments match just fine 
(all three get highlighted).


So... why is SA hitting only on the final comment, and ignoring the 
first two?  (I tried using a meta rule that popped if this hit more 
than once, and that meta rule did not pop.  SA debug output shows 
only this one comment hitting, not the other two.)  If my regex is 
fine and I've got tflags=multiple, what's preventing the other 
comments from hitting?


Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-11 Thread Benny Pedersen

Amir 'CG' Caspi skrev den 2013-08-11 10:22:


http://pastebin.com/VCtvzjzV



Content analysis details:   (10.9 points, 5.0 required)

 pts rule name  description
 -- 
--

-0.0 RCVD_IN_MSPIKE_H3  RBL: Good reputation (+3)
[5.39.218.213 listed in wl.mailspike.net]
 0.1 RELAY_NL   Relayed through NL
 0.5 MSG_ID_INSTAFILE_BIZ   spamming instafile.biz in message id
 0.5 STARS_ON_FORTY_FIVEURI: contains 5 chars url at end
 0.1 STARS_ON_FORTY_FOORURI: contains 4 chars url at end
 0.1 HTML_ERROR_TAGS_X_HTML RAW: error x-html not found on w3.org
 2.4 RAZOR2_CF_RANGE_E8_51_100 Razor2 gives engine 8 confidence level
above 50%
[cf: 100]
 0.4 RAZOR2_CF_RANGE_51_100 Razor2 gives confidence level above 50%
[cf: 100]
 1.7 RAZOR2_CHECK   Listed in Razor2 (http://razor.sf.net/)
-0.0 RCVD_IN_MSPIKE_WL  Mailspike good senders
 1.8 LONGWORDS  Long string of long words
 2.0 MIME_NO_TEXT   No (properly identified) text body parts
 1.3 SAGREY Adds score to spam from first-time senders


i created MSG_ID_INSTAFILE_BIZ and HTML_ERROR_TAGS_X_HTML , but even 
without this rules its spam


Re: New spam rule for specific content

2013-08-11 Thread Amir Caspi
On Aug 11, 2013, at 9:10 AM, Benny Pedersen m...@junc.eu wrote:

 i created MSG_ID_INSTAFILE_BIZ and HTML_ERROR_TAGS_X_HTML , but even without 
 this rules its spam

It is NOW, it was not when it was originally processed, as you can see from the 
SA headers included in the pastebin. If you read the messages I sent earlier, 
the network tests did not all hit because the spam was too young (had not yet 
been reported to all the services). LONGWORDS also did not hit for some reason, 
see the second email I sent regarding this (the test seems to not work properly 
on MIME content). Without these, and because this is an image-based spam that 
evades Bayes, the message did not pass the spam threshold originally, even 
though it does now.

My question is not whether this is spam. My question is why the new 
HTML_COMMENT_GIBBERISH rule only got one hit on the third comment when it 
should have hit all three comments...

Thanks.

--- Amir

Re: New spam rule for specific content

2013-08-11 Thread Amir 'CG' Caspi

At 2:22 AM -0600 08/11/2013, Amir 'CG' Caspi wrote:
My regex is valid and appropriate for those comments... I tested it 
at regexpal.com, which shows that all three comments match just fine 
(all three get highlighted).


So... why is SA hitting only on the final comment, and ignoring the first two?


Further confusion.  Received another of these types of spam today:

http://pastebin.com/YywcFkui

My new HTML_COMMENT_GIBBERISH rule didn't hit on this one at all. 
Running the email through regexpal.com shows that the regex _DOES_ 
hit the comment.  Why is this failing in SA even though it works in 
other environments?  Is there something that Perl doesn't like about 
my regex syntax but that works fine in JavaScript?


Whatever is causing this to fail is probably the same thing causing 
only the single (versus triple) hit on the previous example.


Your help in debugging would be greatly appreciated...

Thanks!

--- Amir


Re: New spam rule for specific content

2013-08-11 Thread Alex
Hi,

 Further confusion.  Received another of these types of spam today:

 http://pastebin.com/YywcFkui

 My new HTML_COMMENT_GIBBERISH rule didn't hit on this one at all. Running

Can you post this rule again so we can investigate?

How do you find the SPAMMY_URI_PATTERNS rule is performing? It seems
very prone to FPs.

Why is there no BAYES score?

Are you using sqlgrey? If not, it's incredible and you should try it.

Regards,
Alex


Re: New spam rule for specific content

2013-08-11 Thread Amir 'CG' Caspi

At 9:31 PM -0400 08/11/2013, Alex wrote:

Can you post this rule again so we can investigate?


# HTML comment gibberish
# Looks for sequence of 100 or more words (alphanum + punct 
separated by whitespace) within HTML comment

rawbody HTML_COMMENT_GIBBERISH  /!--\s*(?:[\w'?!.:;-]+\s+){100,}\s*--/im
describe HTML_COMMENT_GIBBERISH lots of spammy text in HTML comment
score HTML_COMMENT_GIBBERISH0.001

regexpal says my rule matches the comment.  SA doesn't agree.


How do you find the SPAMMY_URI_PATTERNS rule is performing? It seems
very prone to FPs.


It's performing quite well for me... I haven't seen any FPs on it. 
The patterns are based on specific spam templates... one looks for 
/outl and /outi URIs, the other is /land/ + /unsub/ + /report/ ... 
these URIs have to occur in combination.  You are correct that it has 
the potential for FPs but I haven't seen any so far.



Why is there no BAYES score?


I ran this test through the root account which does not have a Bayes 
DB, so there's no Bayes score.  There was a Bayes score on the 
original email, which was Bayes50 just like every other one of these 
types of spams (no real text, just a spammy image which SA isn't 
decoding).



Are you using sqlgrey? If not, it's incredible and you should try it.


I have not implemented any sort of greylisting yet.  I can't use 
sqlgrey because I don't use postfix... my server runs sendmail.  I'm 
sure there are some good sendmail-compatible greylisters but I 
haven't tried them yet... I'm a bit worried about legitimate email 
getting bounced.  I'm sure I'll get to it in due course, though...


Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-11 Thread John Hardin

On Sun, 11 Aug 2013, Amir 'CG' Caspi wrote:


At 2:22 AM -0600 08/11/2013, Amir 'CG' Caspi wrote:
My regex is valid and appropriate for those comments... I tested it at 
regexpal.com, which shows that all three comments match just fine (all 
three get highlighted).


So... why is SA hitting only on the final comment, and ignoring the first 
two?


Further confusion.  Received another of these types of spam today:

http://pastebin.com/YywcFkui

My new HTML_COMMENT_GIBBERISH rule didn't hit on this one at all.


Thanks for the samples, and apologies for the tardy reply.

A COMMENT_GIBBERISH rule has been in my sandbox for a while now, but it is 
not performing well in masscheck.


I broadened it a bit per your samples and it hits all of them now. We'll 
see if this change improves the masscheck performance. I'm also going to 
make FP-avoidance changes that should also help.


Running the email through regexpal.com shows that the regex _DOES_ hit 
the comment.  Why is this failing in SA even though it works in other 
environments?  Is there something that Perl doesn't like about my regex 
syntax but that works fine in JavaScript?


I haven't tested your rule yet, but I have a comment: you are trying a bit 
too hard. Don't worry about matching all the way to the end of the 
comment. You don't care about gibberish past the first 100 words. Just 
make sure that the rule does not match the -- comment-end token, and stop 
at 100 matched words. Past that it doesn't matter.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  The fetters imposed on liberty at home have ever been forged out
  of the weapons provided for defense against real, pretended, or
  imaginary dangers from abroad.   -- James Madison, 1799
---
 4 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-11 Thread Amir 'CG' Caspi

At 6:56 PM -0700 08/11/2013, John Hardin wrote:

I'm also going to make FP-avoidance changes that should also help.


Care to share? =)


Just make sure that the rule does not match the -- comment-end token


I tried doing that and it caused SA to hang... couldn't figure out 
why the regex wasn't working, but for whatever reason, it wasn't.  I 
figured it was easier to just match the entire comment.
	Is there any particular reason to NOT match the entire 
comment?  That is, does it save resources (CPU, RAM, etc.) to match 
only partial content?


Note that you do want to allow HTML tags within the comment... my 
rule doesn't actually allow that, but I've seen spams with HTML tags 
(mostly p and div) in the comments... we don't want to exclude 
those.


Care to post your updated rule?

Either way, I would still love to know why my rule isn't hitting on this...

Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-11 Thread John Hardin

On Sun, 11 Aug 2013, Amir 'CG' Caspi wrote:


At 9:31 PM -0400 08/11/2013, Alex wrote:

Are you using sqlgrey? If not, it's incredible and you should try it.


I have not implemented any sort of greylisting yet.  I can't use sqlgrey 
because I don't use postfix... my server runs sendmail.  I'm sure there are 
some good sendmail-compatible greylisters but I haven't tried them yet...


milter-greylist is what I use, it seems to do the job, and it does reduce 
the spam volume.



I'm a bit worried about legitimate email getting bounced.


The only problem would be with a sending MTA that either is badly 
misconfigured or cannot properly deal with a tempfail result and either 
bounces the message as undeliverable or (worse) quietly drops it.


Sadly there are some major players with this problem who are apparently 
uninterested in fixing their systems. I suggest you do a bit of research 
on whitelists for greylisting before implementation. You would also 
probably want to whitelist known regular correspondents.


There's also the need to set your users' expectations. They should be 
trained that email is *not*, and is not intended to be, instantaneous.



--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  The fetters imposed on liberty at home have ever been forged out
  of the weapons provided for defense against real, pretended, or
  imaginary dangers from abroad.   -- James Madison, 1799
---
 4 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-11 Thread John Hardin

On Sun, 11 Aug 2013, Amir 'CG' Caspi wrote:


At 6:56 PM -0700 08/11/2013, John Hardin wrote:

I'm also going to make FP-avoidance changes that should also help.


Care to share? =)


Everything is publicly visible in my sandbox:
http://svn.apache.org/viewvc/spamassassin/trunk/rulesrc/sandbox/jhardin/

The results for the rule set are here:
http://ruleqa.spamassassin.org/detail?rule=%2FCOMMENT_GIBBERISHsrcpath=jhardin


Just make sure that the rule does not match the -- comment-end token


I tried doing that and it caused SA to hang... couldn't figure out why the 
regex wasn't working, but for whatever reason, it wasn't.


The unbounded matches you're using probably caused the RE engine to get 
stuck backing off and retrying. REs are by default greedy, they try to 
match as much as possible.


In general it is a *VERY BAD* idea to use * or + in SA REs; they are 
only really safe in rules that process data that is already limited in 
size, like uri rules or header rules that look at a specific header. Make 
it a habit to use bounded matches, {0,n} rather than * and {1,n} rather 
than +. The upper bound of {n} will limit how much the engine will back 
off and retry.


Our rules are similar, take a look at what I have in the sandbox.


I figured it was easier to just match the entire comment.
	Is there any particular reason to NOT match the entire comment?  That 
is, does it save resources (CPU, RAM, etc.) to match only partial content?


It does. The less text you match beyond what you need to, the less 
processing is performed. Nothing is done with the matched text, so the 
extra work done matching all the way to the end of the comment is wasted.


Note that you do want to allow HTML tags within the comment... my rule 
doesn't actually allow that, but I've seen spams with HTML tags (mostly p 
and div) in the comments... we don't want to exclude those.


Yuck. Can you pastbin spamples, if you still have them?


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Efficiency can magnify good, but it magnifies evil just as well.
  So, we should not be surprised to find that modern electronic
  communication magnifies stupidity as *efficiently* as it magnifies
  intelligence.   -- Robert A. Matern
---
 4 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-11 Thread Amir 'CG' Caspi

At 7:20 PM -0700 08/11/2013, John Hardin wrote:
The unbounded matches you're using probably caused the RE engine to 
get stuck backing off and retrying.


That's what I figured.  That's why I changed things to the current 
version, which is bounded by the end-tag of the comment.  My 
current version doesn't take long to run.




Yuck. Can you pastbin spamples, if you still have them?


Here's one that comes to mind:

http://pastebin.com/zVEH2h02

I have a couple of others but they look like they're from the same 
template, so I don't think it's useful to post.


--- Amir


Re: New spam rule for specific content

2013-08-11 Thread John Hardin

On Sun, 11 Aug 2013, Amir 'CG' Caspi wrote:


At 7:20 PM -0700 08/11/2013, John Hardin wrote:

Yuck. Can you pastbin spamples, if you still have them?


Here's one that comes to mind:

http://pastebin.com/zVEH2h02


That's going to be problematic as the comment isn't gibberish, it's a 
bunch of properly-formed sentences.


However, I may be taking too-conservative a stance here. It's possible 
that, while HTML comments can appear in ham, *long* HTML comments won't, 
and the fact that we're looking for long blocks of comment text is enough 
safety.


I'll play around with that sample and see what happens.

--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  [People] are socialists because they are blinded by envy and
  ignorance.-- economist Ludwig von Mises (1881-1973)
---
 4 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-11 Thread Amir 'CG' Caspi

At 8:23 PM -0700 08/11/2013, John Hardin wrote:
However, I may be taking too-conservative a stance here. It's 
possible that, while HTML comments can appear in ham, *long* HTML 
comments won't, and the fact that we're looking for long blocks of 
comment text is enough safety.


That's why feeling.  You'll notice my rule is dumb: it's simply 
looking for a bunch of stuff in a comment.  My main feeling is that 
if anyone is sending HTML email with LOTS of stuff commented out, 
that email is almost certainly spam.  Ham HTML email would probably 
be done with more care.


Yes, there's the chance for FPs, if some company decides to send a 
legitimate (ham, opt-in, etc.) HTML email from a badly-written 
template where the designer was a lazy bum and left giant 
commented-out sections... but would you really want such an email 
anyway? ;-)


Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-10 Thread Amir 'CG' Caspi

At 10:41 AM -0700 08/09/2013, John Hardin wrote:

Can you provide a spample or two?


Sure.
http://pastebin.com/VfSCB7fw
http://pastebin.com/VCtvzjzV

Note the outl and outi links near the very bottom.  The actual 
domains used in these URIs vary... they used to be .pw, but recently 
most have been .biz (though I've also seen some .mobi and I think 
some .tv and even some .us).


Note that both of these hit BAYES_50... and that's pretty common for 
these spams.  For whatever reason, I don't know why, they seem to 
only hit BAYES_50 and very rarely get higher scores (occasionally 
they will get lower scores, too).  Perhaps it's because most of the 
spam is actually in the embedded image, rather than in rendered 
text...


These are also great examples of the HTML comment gibberish that 
pervades all of these spams.  If you have time, it would be great if 
you could adapt your STYLE_GIBBERISH rules to catch HTML comment 
gibberish.  (Presumably, you'd want to make sure the gibberish is 
sufficiently long, too.)


They can be added but unless such spams appear in the masscheck 
corpora the rules won't be scored and distributed.


No idea if they're in the masscheck corpora... but I and my users 
have been getting them for months.  I imagine they're relatively 
widespread...


Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-10 Thread Amir 'CG' Caspi

At 10:41 AM -0700 08/09/2013, John Hardin wrote:

Can you provide a spample or two?


Looks like a similar spam method has come out in recent weeks (since 
Jul 30, it seems) that uses slightly different footers... example is 
here:


http://pastebin.com/QCmSPzwG

Although running SA on this spam _NOW_ yields a high score beyond the 
spam threshold, this is almost entirely because additional network 
tests are now hitting (extra RBLs + Razor).  This was not the case 
when the spam was first processed... looks like I was one of the 
earlier recipients.


For this type, looks like a good match would be on the combo of 
/land/ + /unsub/ + /report/ ... I have modified my rule from 
yesterday as follows:


# Spammy URI patterns
uri __OUTL_URI  /\/outl\b/
uri __OUTI_URI  /\/outi\b/
uri __LAND_URI  /\/land\//
uri __UNSUB_URI /\/unsub\//
uri __REPORT_URI/\/report\//
meta SPAMMY_URI_PATTERNS((__OUTL_URI  __OUTI_URI) || 
(__LAND_URI  __UNSUB_URI  __REPORT_URI))

describe SPAMMY_URI_PATTERNSlink combos match highly spammy template
score SPAMMY_URI_PATTERNS   3

This modification hits both types of templates.  I will very likely 
be adding further spammy patterns to this rule over time.  I'll 
keep the list posted if I find some other good ones.



It looks like both this and the previous type of spam are bypassing 
Bayes by embedding images and using no rendered text.  Well, not NO 
text, but very little, mostly a successful delivery message and the 
unsub/report links.  That is, Bayes sees absolutely no spammy text, 
just the image which it cannot decode as spammy.


Are there any rules which can hit on only embedded images with very 
little text ??  Not entirely sure how to capture this since it's 
difficult to determine what is not much text and there is certainly 
the potential for FPs that way (for example, anyone in the design 
field sending images to clients without much text, etc.)...


But, these types of spams are bypassing SA consistently, to the tune 
of tens per day per user.  I would really love a way to stop them 
besides hardcoding a rule based on their link syntax, which can be 
easily changed during the next iteration of their spam template.


(The HTML comment gibberish rule would be a big step here, since 
that's one of the few things that would distinguish this from ham... 
unlikely that a real person would embed tens of KB of comment 
gibberish.)


Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-10 Thread John Hardin

On Sat, 10 Aug 2013, Amir 'CG' Caspi wrote:

It looks like both this and the previous type of spam are bypassing Bayes by 
embedding images and using no rendered text.  Well, not NO text, but very 
little, mostly a successful delivery message and the unsub/report links. 
That is, Bayes sees absolutely no spammy text, just the image which it 
cannot decode as spammy.


Perhaps it's time to bring FuzzyOCR up-to-date...?


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  The social contract exists so that everyone doesn't have to squat
  in the dust holding a spear to protect his woman and his meat all
  day every day. It does not exist so that the government can take
  your spear, your meat, and your woman because it knows better what
  to do with them.   -- Dagny @ Ace of Spades
---
 5 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-10 Thread Amir 'CG' Caspi

At 2:17 PM -0700 08/10/2013, John Hardin wrote:


Perhaps it's time to bring FuzzyOCR up-to-date...?


Is this something I need to manually update or something that needs 
updating in the SA distribution?


Thanks.

--- Amir


Re: New spam rule for specific content

2013-08-10 Thread John Hardin

On Sat, 10 Aug 2013, Amir 'CG' Caspi wrote:


At 2:17 PM -0700 08/10/2013, John Hardin wrote:


Perhaps it's time to bring FuzzyOCR up-to-date...?


Is this something I need to manually update or something that needs updating 
in the SA distribution?


FuzzyOCR was a SA plugin a few years back. It would pass images through 
OCR and, IIRC, pull words out of them into the generated body that SA 
scans.


Spammers moved away from putting their spams into images, so it fell out 
of use, and I don't think it works with the current release of SA. Also, 
Passing all attached images through OCR is a fairly heavy-weight process.


Now spammers seem to be moving back towards image spams, at least to a 
degree.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  ...in the 2nd amendment the right to arms clause means you have
  the right to choose how many arms you want, and the militia clause
  means that Congress can punish you if the answer is none.
-- David Hardy, 2nd Amendment scholar
---
 5 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-09 Thread John Hardin

On Fri, 9 Aug 2013, Amir 'CG' Caspi wrote:

	A number of my users have been receiving spam formatted in a very 
specific way which seems to very often miss Bayes...


Can you provide a spample or two?


I recommend this rule be added to the general distribution.


They can be added but unless such spams appear in the masscheck corpora 
the rules won't be scored and distributed.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  The first time I saw a bagpipe, I thought the player was torturing
  an octopus. I was amazed they could scream so loudly.
-- cat_herder_5263 on Y! SCOX
---
 6 days until the 68th anniversary of the end of World War II


Re: New spam rule for specific content

2013-08-09 Thread RW
On Fri, 9 Aug 2013 11:19:08 -0600
Amir 'CG' Caspi wrote:

   A number of my users have been receiving spam formatted in a 
 very specific way which seems to very often miss Bayes... I don't 
 know why, whether it's because of the HTML gibberish flooding Bayes 
 with useless tokens (to reduce the relative strength of the spammy 
 tokens), or if it's just the specific content isn't sufficiently 
 spammy (or has sufficient ham to balance) to pop.

BAYES works on rendered text it doesn't see the HTML.


 (Like many other users here, I've also increased the Bayes scores for 
 Bayes99, and created a Bayes999 with even higher scoring... it might 
 be time to add that to the general distribution, too.)

Do you actually get a significant amount of ham between 0.99 and 0.999?
Personally I only get 1 in 1000 above 0.55, and nothing above 0.65.


Re: New spam rule for specific content

2013-08-09 Thread Amir 'CG' Caspi
On Fri, August 9, 2013 1:01 pm, RW wrote:
 BAYES works on rendered text it doesn't see the HTML.

Hmmm.  It doesn't see HTML comments, which would appear in rendered HTML
source even though they are invisible?  OK, in that case, I have NO idea
why the spam isn't hitting Bayes, because it looks pretty damn spammy to
me.  I wonder if it's the heavy use of images, but I don't know.

 Do you actually get a significant amount of ham between 0.99 and 0.999?
 Personally I only get 1 in 1000 above 0.55, and nothing above 0.65.

Ham, absolutely not.  So yes, I suppose I could just treat all Bayes99 as
if it were Bayes999 and score it more highly than I do.  Right now I have
Bayes99 at 4, Bayes999 at 4.5.  I could eliminate Bayes999 and make
Bayes99 score 4.5... but I do worry a little bit about FPs, even though I
guess I shoudn't, statistically speaking.

On the other hand, one could consider making Bayes999 a poison pill. 
Generally spam will only rank there if you've learned something nearly
identical to it.  At that point, perhaps it might be worth just scoring it
with 5 or higher (assuming your threshold is 5, as mine is).

--- Amir



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-08-05 Thread Jason L Tibbitts III
 DS == Dan Schaefer d...@performanceadmin.com writes:

DS I'm glad to see this SPAM traffic has come to a halt. At least on my
DS mail server...

Yes, I haven't seen any of those spams since the morning of the 31st.
My servers were rejecting them like mad right up until that point in
time (10:30CDT), and then nothing.

 - J


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-08-05 Thread Michelle Konzack
Good morning *,

Am 2009-08-04 13:51:24, schrieb Jason L Tibbitts III:
  DS == Dan Schaefer d...@performanceadmin.com writes:
 
 DS I'm glad to see this SPAM traffic has come to a halt. At least on my
 DS mail server...
 
 Yes, I haven't seen any of those spams since the morning of the 31st.
 My servers were rejecting them like mad right up until that point in
 time (10:30CDT), and then nothing.

I have seen exactly the same, I was hit by more then 200.000  spams  per
day of this kind and had a relative  high  CPU  load  (4)  on  my  five
servers Sun Fire X4100M2 and it was more or less gone from one hour to
another...

Thanks, Greetings and nice Day/Evening
Michelle Konzack
Systemadministrator

Tamay Dogan Network
Debian GNU/Linux Consultant

-- 
Linux-User #280138 with the Linux Counter, http://counter.li.org/
# Debian GNU/Linux Consultant #
http://www.tamay-dogan.net/ Michelle Konzack
http://www.can4linux.org/   c/o Vertriebsp. KabelBW
http://www.flexray4linux.org/   Blumenstrasse 2
Jabber linux4miche...@jabber.ccc.de   77694 Kehl/Germany
IRC #Debian (irc.icq.com) Tel. DE: +49 177 9351947
ICQ #328449886Tel. FR: +33  6  61925193


signature.pgp
Description: Digital signature


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-08-04 Thread Dan Schaefer
I'm glad to see this SPAM traffic has come to a halt. At least on my 
mail server...


--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-08-04 Thread Michelle Konzack
Hi Dan and *,

Am 2009-08-04 14:37:46, schrieb Dan Schaefer:
 I'm glad to see this SPAM traffic has come to a halt. At least on my  
 mail server...

They have seen, the out spamassassin is working verry efficient.  I  get
only one or two spams per day...  which are catched by SA of course.

Thanks, Greetings and nice Day/Evening
Michelle Konzack
Systemadministrator
Tamay Dogan Network
Debian GNU/Linux Consultant

-- 
Linux-User #280138 with the Linux Counter, http://counter.li.org/
# Debian GNU/Linux Consultant #
http://www.tamay-dogan.net/ Michelle Konzack
http://www.can4linux.org/   c/o Vertriebsp. KabelBW
http://www.flexray4linux.org/   Blumenstrasse 2
Jabber linux4miche...@jabber.ccc.de   77694 Kehl/Germany
IRC #Debian (irc.icq.com) Tel. DE: +49 177 9351947
ICQ #328449886Tel. FR: +33  6  61925193


signature.pgp
Description: Digital signature


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Daniel J McDonald
On Thu, 2009-07-23 at 07:34 +0100, rich...@buzzhost.co.uk wrote:
 It's catching on :-)

this new obfuscation is already caught by AE_MED45, but I can foresee a
variant that might not match...

How about:

body__MED_OB
/\bw{2,3}(?:[[:punct:][:space:]]{1,5}|[[:space:][:punct:]]{1,3}dot[[:space:][:punct:]]{1,3})[[:alpha:]]{0,6}\d{2,6}(?:[[:punct:][:space:]]{1,5}|[[:space:][:punct:]]{1,3}dot[[:space:][:punct:]]{1,3})(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)[[:punct:]]?\b/i
body__MED_NOT_OB/\bw{2,3}\.[[:alpha:]]{0,6}\d{2,6}\.(?:com|net|org)\b/i
metaAE_MED46(__MED_OB  ! __MED_NOT_OB)
describeAE_MED46Shorter rule to catch spam obfuscation
score   AE_MED464.0

-- 
Dan McDonald, CCIE #2495, CISSP# 78281, CNX
www.austinenergy.com


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Dan Schaefer



For those of you that manage these rules,
URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this 
email as spam


I'm up to AE_MED45, so I wouldn't expect AE_MEDS38 and 39 to be 
hitting anything currently.


http://pastebin.com/m40f7cff4

This is not an obfuscated domain.  You can see that it hit two URIBLs 
- JP and WS.  I would have expected it to be in URIBL_BLACK (or at 
least GOLD) as well as Invaluement's URIBL.  There are plenty of 
mechanisms to catch valid URIs - that's not the purpose of the 
obfuscation rules.


And, you still got 15 points - so, what's the problem?

Relax. I don't have a problem. I was just pointing out a potential flaw. 
I was just trying to help out. I just misunderstood the whole blacklist 
thing, that's all.


--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Dan Schaefer



It means that if you were using BL at MTA level your SA might never have seen 
the message at all.

No your rule would not be overlooked 'because the site is in a blacklist' 
*unless* you were using the BL in your MTA and rejected the transaction from a 
blacklisted IP address and, thus, never submitted it to SA at all.

  
If this is the case, then why does my email have the X-* headers in it? 
I have nothing in my postfix header_checks to discard the BL rules. Does 
anyone have a detailed flow chart of SA/postfix setup and describes 
blacklisting? Or even a webpage describing the process?


--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Bowie Bailey

Dan Schaefer wrote:


It means that if you were using BL at MTA level your SA might never 
have seen the message at all.


No your rule would not be overlooked 'because the site is in a 
blacklist' *unless* you were using the BL in your MTA and rejected 
the transaction from a blacklisted IP address and, thus, never 
submitted it to SA at all.


  
If this is the case, then why does my email have the X-* headers in 
it? I have nothing in my postfix header_checks to discard the BL 
rules. Does anyone have a detailed flow chart of SA/postfix setup and 
describes blacklisting? Or even a webpage describing the process?


It's very simple with Postfix or any other MTA.

1) Connection request comes to Postfix.
2) Postfix checks the sending server against its blacklists.  If it 
matches, the mail is refused.
3) Postfix checks its normal rules and if the sender/recipient/etc is 
ok, the message is accepted.

4) Postfix sends the message to SA.
5) SA scores the message and returns it to Postfix (SA blacklists simply 
score 100 points).
6) Postfix can now deliver, quarantine or delete the message based on 
the score or spam/ham designation returned by SA.


--
Bowie


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Charles Gregory

On Wed, 22 Jul 2009, Dan Schaefer wrote:

For those of you that manage these rules,
URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this 
email as spam

http://pastebin.com/m40f7cff4


The URI is not obfuscated, therefore it triggered the URIBL tests 
properly (and scored 3 additional points from them).


- C


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Martin Gregorie
Dan Schaefer wrote:

 If this is the case, then why does my email have the X-* headers in 
 it? I have nothing in my postfix header_checks to discard the BL 
 rules. Does anyone have a detailed flow chart of SA/postfix setup and 
 describes blacklisting? Or even a webpage describing the process?
 
Are you quite sure that an upstream copy of SA, e.g. in your ISP or at a
sender site that scans for outgoing spam, hasn't already added X-*
headers to the message?


Martin




Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Dan Schaefer



Are you quite sure that an upstream copy of SA, e.g. in your ISP or at a
sender site that scans for outgoing spam, hasn't already added X-*
headers to the message?


Martin

  

No. Is that even possible to track down?

--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Dan Schaefer



 Are you quite sure that an upstream copy of SA, e.g. in your ISP or at
 a sender site that scans for outgoing spam, hasn't already added X-*
 headers to the message?


No. Is that even possible to track down?


There would probably be an X-Spam-Checker-Version header in your 
inbound mail stream.


X-Spam-Checker-Version: SpamAssassin 3.2.5 (2008-06-10) on 
pony.performanceadmin.com


That is my server.

--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread John Hardin

On Thu, 23 Jul 2009, Dan Schaefer wrote:


 Are you quite sure that an upstream copy of SA, e.g. in your ISP or at
 a sender site that scans for outgoing spam, hasn't already added X-*
 headers to the message?


No. Is that even possible to track down?


There would probably be an X-Spam-Checker-Version header in your inbound 
mail stream.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Perfect Security and Absolute Safety are unattainable; beware
  those who would try to sell them to you, regardless of the cost,
  for they are trying to sell you your own slavery.
---
 12 days since a sunspot last seen - EPA blames CO2 emissions


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Martin Gregorie
On Thu, 2009-07-23 at 12:25 -0400, Dan Schaefer wrote:
  Are you quite sure that an upstream copy of SA, e.g. in your ISP or at a
  sender site that scans for outgoing spam, hasn't already added X-*
  headers to the message?
 
 
  Martin
 

 No. Is that even possible to track down?
 
Sure - look at any incoming message's headers to see if there are any
that didn't come from your copy of SA. Each set has a
X-spam-checker-version header that gives the name of the SA host that
added that header set. If that's a possibility, just make sure your
filter ignores header sets that aren't yours. AFAIK your SA header set
it always the first in the message headers.


Martin
 



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread John Hardin

On Thu, 23 Jul 2009, Dan Schaefer wrote:

Are you quite sure that an upstream copy of SA, e.g. in your ISP 
or at a sender site that scans for outgoing spam, hasn't already 
added X-* headers to the message?
 
  No. Is that even possible to track down?


 There would probably be an X-Spam-Checker-Version header in your
 inbound mail stream.


X-Spam-Checker-Version: SpamAssassin 3.2.5 (2008-06-10) on 
pony.performanceadmin.com


That is my server.


You'd have to check for that _before_ your local SA got a crack at the 
message. Whether you can grab a copy of mail before SA depends on your 
glue.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Win95: Where do you want to go today?
  Vista: Where will Microsoft allow you to go today?
---
 12 days since a sunspot last seen - EPA blames CO2 emissions


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-23 Thread Kevin Parris
(apologies for top posting, but the email software here does not really do 
quoting in a way that works out well otherwise)

If your mail contains SpamAssassin headers then it was (obviously) processed 
through SpamAssassin.  Just because you have BL checks in your MTA does not 
necessarily mean that all spam items will be blocked at that level.  Lots of 
spam can pass some BL checks and then be scored high as the result of other 
things.

My comments were not meant to say that BL checks stop spam.  I was responding 
specifically to your inquiry about a rule being 'overlooked' if there happened 
to be a message it would hit that also had something in it that would hit a 
blacklist too.  I think you're reading too much complexity into things.  Or 
maybe not enough.

The basic idea is something like this:

 a) You have some stuff specified for Postfix to do, it starts doing those 
things, and if it gets through them (without deciding to reject the message) to 
the point where you specify a call to SA, then it passes the item to SA for 
scoring.

 b)  SA applies the rules (which usually include querying various blacklists 
based on things found within the message) and tallies up the score, then it 
gives the results to whatever asked it to analyze the message.

 c)  Then whatever that was (in your case, Postfix) looks at the results and 
decides what to do next, based on what you specified for it.

SpamAssassin does not block mail.  SpamAssassin analyzes a message and assigns 
a score.   Mail handlers reject/quarantine/discard/deliver mail.  SpamAssassin 
is not a mail handler.

If you don't understand the effects of entries in your Postfix configuration, 
you probably will get better assistance in a Postfix-specific forum.

 Dan Schaefer d...@performanceadmin.com 07/23/09 10:22 AM 

 It means that if you were using BL at MTA level your SA might never have seen 
 the message at all.

 No your rule would not be overlooked 'because the site is in a blacklist' 
 *unless* you were using the BL in your MTA and rejected the transaction from 
 a blacklisted IP address and, thus, never submitted it to SA at all.

   
If this is the case, then why does my email have the X-* headers in it? I have 
nothing in my postfix header_checks to discard the BL rules. Does anyone have a 
detailed flow chart of SA/postfix setup and describes blacklisting? Or even a 
webpage describing the process?




Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread twofers

Charles,
Because we CAN'T.
My point exactly. No matter what, with the current system of internet email, 
SPAM will never be stopped or filtered out completely. A completely new concept 
of verifying internet email would be required for that and unfortunately, that 
will never happen simply because It's all about the money and as far as this 
is concerned, it generates a revenue stream, it generates new technologies 
concepts and tax revenue. The governments not going to stifle that, the 
government is going to allow the industry to regulate itself, one way or the 
otheras long as it generates revenues and taxes. It's simply Capitalism at 
work.
SPAM email will never be completely eliminated, it will only, ever just be 
minimized based on the current system.
False positives, a fact of filtering that beckon for refinement, for tweaking 
and for precise detailing of the filters rules.
Even our Good Ideas are not fallible. Without the SPAMMERS knowledge of the 
rules, they are static and complacent. With the SPAMMERS knowledge of the 
rules, they are dynamic, correctable, upgradeable and ever so more restrictive 
and precise over time, designed to extract precisely a balance between the 
legitimate and non legitimate.
We can't fine-tune anything if we do not have a means of measuring our 
requirements. Eventually the SA rules will refine themselves to a precision 
that will be virtually impregnable by SPAMMERS. The sooner that happens the 
better and it will happen sooner as the SPAMMERS show us their means and they 
are adapted to our requirements.
I'm sure the powers that be who make SA public as it is did so for a reason, 
or were not expressly concerned over it's exposure.
There is nothing the SPAMMERS can send that can't be filtered to a high degree. 
It's not about eliminating, it's about minimizing.
On Tue, 21 Jul 2009, twofers wrote:
  so why not let them show us what they've got, show us where we need to 
 make adjustments and corrections and in turn we will continue to refine our 
 process, ever so more, squeezing them out...inch by inch.  

Because we CAN'T. While the spammers are free to try ANY obfuscation or 
filter-dodging technique imaginable, we are always constrained to avoid false 
positives. So any time we share our 'good ideas' with them, they come closer to 
their 'goal' of finding the 'perfect' way to spam that we cannot filter...

And as a side note, I've noticed that I might have a rule in place, like my 
original, simple 'shopXX' rule, and it worked for me for a couple of weeks, 
until people started posting rules for it here. Then the more-complex 
obfuscations started
And we started correcting and upgrading and fine tuning our rules to meet those 
new requirements...all the while, the SPAMMERS were shooting themselves in the 
foot as far as their click rates were concernedclick rates their customers 
use to validate their expenses for that form of advertisement
I would venture to say that the SPAMMERS were grasping or otherwise just 
plain teasing as their return on investment was going straight into the 
toilet.Wes


  

Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread Benny Pedersen

On Wed, July 22, 2009 13:16, twofers wrote:
 Because we CAN'T.

Obama says yes we can :)

 My point exactly. No matter what, with the current system of internet email,

just becurse main stream spammers is so clueless that thay start using 
recipient equal to sender evelope says thay newer got used
to spf ?

 SPAM will never be stopped or filtered out completely.

wroung

 A completely new concept of verifying internet email would be required for 
 that and unfortunately,

as in dkim/spf no ?

 that will never happen simply because It's all about the money

spf and dkim is gpl

 and as far as this is concerned, it generates a revenue stream,

where ?

 it generates new technologies concepts and tax revenue.

where ?

 The governments not going to stifle that,

do governments use spf/dkim ?

 the government is going to allow the industry to regulate itself,

good :)

 one way or the otheras long as it generates revenues and taxes.

who gets the money ?

 It's simply Capitalism at work.

just stop paying

 SPAM email will never be completely eliminated,

wroung, we can close our email box also

 it will only, ever just be minimized based on the current system.

ah you now admit we can win ?

 False positives, a fact of filtering that beckon for refinement,

imho there is none if recipient add friends to his address book, and that 
addressbook is dumped to whitelist_auth in sa

 for tweaking and for precise detailing of the filters rules.

we already have to many rules in sa imho, it turns down to sender is known or 
not :/

 Even our Good Ideas are not fallible. Without the SPAMMERS knowledge of the 
 rules,

start thinking more on what spammers cant do for us might be the route to stop 
spammers for just get a bunch of new meds domains
with numbers in end, start using url that whitelist, but only apply white if 
there is no other url !

 they are static and complacent. With the SPAMMERS knowledge of the rules,

you belive that spammers using sa to test the spam runs ?, if yes why do i see 
80% spam mails get rejected with spf testing alone ?

 they are dynamic, correctable, upgradeable and ever so more restrictive and 
 precise over time,

well its maybe currect that clever spammers can find another way of being 
clueless when using sa to test there spam goals, but it
will not make most sa installs not detect it as spam, bayes can cougt anything

 designed to extract precisely a balance between the legitimate and non 
 legitimate.

bayes working

 We can't fine-tune anything if we do not have a means of measuring our 
 requirements.

currect, but if we make sure sender is not forged, and whitelist known senders, 
this is a start, if this is not done we have more
complex work to do before its possible to stop spam

also why there is so much new rules to stop new spam, its endless :/

 Eventually the SA rules will refine themselves to a precision that will be 
 virtually impregnable by SPAMMERS.

dkim is nice, but it creates lots of load to test this in mta since we need to 
recieve whole email before dkim testing can be
tested :/

thats why is say go to spf

 The sooner that happens the better and it will happen sooner as the SPAMMERS 
 show us their means and they are adapted
 to our requirements. I'm sure the powers that be who make SA public as it 
 is did so for a reason,

its made public so any antispam users can commit rules to fight spammers where 
it hurts :)

 or were not expressly concerned over it's exposure.

maybe

 There is nothing the SPAMMERS can send that can't be filtered to a high 
 degree.

exactly

 It's not about eliminating, it's about minimizing.

agree

-- 
xpoint



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread Dan Schaefer

For those of you that manage these rules,
URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this email as 
spam

http://pastebin.com/m40f7cff4


--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread Benny Pedersen

On Wed, July 22, 2009 21:39, Dan Schaefer wrote:
 For those of you that manage these rules,
 URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this email as 
 spam
 http://pastebin.com/m40f7cff4

reject it with rbl testing in mta, and its found in blacklist, reason it not 
found in obfu is that its not obfu :)

-- 
xpoint



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread Dan Schaefer

Benny Pedersen wrote:

On Wed, July 22, 2009 21:39, Dan Schaefer wrote:
  

For those of you that manage these rules,
URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this email as 
spam
http://pastebin.com/m40f7cff4



reject it with rbl testing in mta, and its found in blacklist, reason it not 
found in obfu is that its not obfu :)

  
Does this mean that if I have a custom rule to search for exactly the 
via site, my rule will be overlooked because the site is in a blacklist?


--
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread Benny Pedersen

On Wed, July 22, 2009 21:56, Dan Schaefer wrote:
 Does this mean that if I have a custom rule to search for exactly the
 via site, my rule will be overlooked because the site is in a blacklist?

what problem ?

-- 
xpoint



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread McDonald, Dan
From: Dan Schaefer [mailto:d...@performanceadmin.com]

For those of you that manage these rules,
URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this email as 
spam

I'm up to AE_MED45, so I wouldn't expect AE_MEDS38 and 39 to be hitting 
anything currently.

http://pastebin.com/m40f7cff4

This is not an obfuscated domain.  You can see that it hit two URIBLs - JP and 
WS.  I would have expected it to be in URIBL_BLACK (or at least GOLD) as well 
as Invaluement's URIBL.  There are plenty of mechanisms to catch valid URIs - 
that's not the purpose of the obfuscation rules.

And, you still got 15 points - so, what's the problem?

--
Dan




Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-22 Thread Kevin Parris
It means that if you were using BL at MTA level your SA might never have seen 
the message at all.

No your rule would not be overlooked 'because the site is in a blacklist' 
*unless* you were using the BL in your MTA and rejected the transaction from a 
blacklisted IP address and, thus, never submitted it to SA at all.

And those rules did not hit on the message because there isn't anything in 
there that they are designed to find.  It does not represent another variation 
on the theme. But since there is a lot of other stuff that other rules did hit 
on, why are you worrying so much about just these few?

 Dan Schaefer d...@performanceadmin.com 07/22/09 3:56 PM 
Benny Pedersen wrote:
 On Wed, July 22, 2009 21:39, Dan Schaefer wrote:
   
 For those of you that manage these rules,
 URI_OBFU_X9_WS, URI_OBFU_WWW, AE_MEDS38, AE_MEDS39 did not mark this email 
 as spam
 http://pastebin.com/m40f7cff4 
 

 reject it with rbl testing in mta, and its found in blacklist, reason it not 
 found in obfu is that its not obfu :)

   
Does this mean that if I have a custom rule to search for exactly the 
via site, my rule will be overlooked because the site is in a blacklist?

-- 
Dan Schaefer
Web Developer/Systems Analyst
Performance Administration Corp.




Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-21 Thread twofers
Charles,
 
Although I understand your reservations, I feel in this case that it's best to 
lay it all out there and give it to them, let them do what they do. In my mind 
it's nothing more than Flushing out the best they can offer and finding the 
loopholes, and closing them up.
 
There are more rules/ways to stop them than they have to defeat the rules and 
scoring process, so why not let them show us what they've got, show us where we 
need to make adjustments and corrections and in turn we will continue to refine 
our process, ever so more, squeezing them out...inch by inch.
 
We will accomplish that goal much quicker if the spammers show us whereall our 
faults lie.
 
Wes
 
On Wed, 15 Jul 2009, MrGibbage wrote:
 I wonder if the spammers are reading this forum.  That seemed awful fast.

I'm sure they do. But I also suspect that they have a simple 'feedback' 
mechanism that let's them know how much of their spew is getting rejected
on their botnets, and when the rejection numbers get too high they try 
something new, and keep trying until the rejection numbers drop again.

Then we fix our rules, the rejections go up, and they look for yet another 
'trick' to get through. They have the advantage of being able to download their 
own copies of spamassassin to 'test' their spew. That's why sometimes you get 
'red herrings' from me on this list when I don't share the full details of a 
rule. Posting it here almost assures that it will get bypassed. They copy the 
rule, then try all sorts of different combinations to bypass it

Now really, the significant factor here is not that any of these obfuscation 
tricks are 'new', but that they are using them to bypass the URIBL rules. I 
strongly urge the spamassassin develpopers to consider ways to 'open up' the 
way that we can specify what SA will 'consider' a URI, or to be able to 
'capture' a value from an obfuscation test, manipulate it into its 'original' 
URI and then 'manually' submit it to the URIBL

Example hypothetical syntax (note that some parentheses are *capturing*):

body FINDURI /(www)(?:obfuscation)(domain)(?:obfuscation)(com|net|org)/i
uribl CHECIT /$1.$2.$3/

Basically, allow a rule to 'capture' one or more 'matches' in Perl variables, 
and then feed them to a subsequent rule (in this case, a manual URIBL lookup). 
This way, the SA developers don't have to hard-code an ever-changing set of 
URI detection rules into the core code, but we can still develop on-the-fly 
rules that can feed a URI to the URIBL tests

I've heard people mention 'plugins'. Could I code one that would be
easily 'modifiable' so that (for example) this morning's '[dot]' trick can be 
quickly added to my plugin? Is there a good working example of a plugin that 
extracts text from a message and feeds it to a URI? I'll work on this!

- C



  

Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-21 Thread Charles Gregory

On Tue, 21 Jul 2009, twofers wrote:
 so why not let them show us what they've got, show us where we 
need to make adjustments and corrections and in turn we will continue to 
refine our process, ever so more, squeezing them out...inch by inch.  


Because we CAN'T. While the spammers are free to try ANY obfuscation or 
filter-dodging technique imaginable, we are always constrained to avoid 
false positives. So any time we share our 'good ideas' with them, they 
come closer to their 'goal' of finding the 'perfect' way to spam that we 
cannot filter...


And as a side note, I've noticed that I might have a rule in place, like 
my original, simple 'shopXX' rule, and it worked for me for a couple of 
weeks, until people started posting rules for it here. Then the 
more-complex obfuscations started


Further to my original post, I haven't read all of today's mail yet, but
I suspec there is not an answer yet to this question, but I wish to 
reiterate it, with a further comment. The comment is that I was looking at 
plugins and noticed that there was one to follow URI's that appear to be
redirects, and 'add' the target URI to the internal list of URI's to be 
run through the URIBL. I tried to look at the script to see if I could 
modify it to my purpose, but just can't figure it out. (sigh)


But it would be a good starting basis for the plugin I am hoping to see.

Original request:
 I strongly urge the spamassassin develpopers to consider ways to 
'open up' the way that we can specify what SA will 'consider' a URI, or 
to be able to 'capture' a value from an obfuscation test, manipulate it 
into its 'original' URI and then 'manually' submit it to the URIBL


Example hypothetical syntax (note that some parentheses are *capturing*):

body FINDURI /(www)(?:obfuscation)(domain)(?:obfuscation)(com|net|org)/i
uribl CHECKIT /$1.$2.$3/

Basically, allow a rule to 'capture' one or more 'matches' in Perl
variables, and then feed them to a subsequent rule (in this case, a manual
URIBL lookup). This way, the SA developers don't have to hard-code an
ever-changing set of URI detection rules into the core code, but we can
still develop on-the-fly rules that can feed a URI to the URIBL tests

I've heard people mention 'plugins'. Could I code one that would be
easily 'modifiable' so that (for example) this morning's '[dot]' trick can
be quickly added to my plugin? Is there a good working example of a plugin
that extracts text from a message and feeds it to a URI? I'll work on this!



- C

Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-21 Thread Karsten Bräckelmann
Sometimes I wished everyone getting involved in heated discussions and
proposals, also would carefully read any post with a related topic...


On Tue, 2009-07-21 at 11:29 -0400, Charles Gregory wrote:
 Further to my original post, I haven't read all of today's mail yet, but

FWIW, neither did I, as I am busy hacking -- and now live. ;)

 Original request:
   I strongly urge the spamassassin develpopers to consider ways to 
  'open up' the way that we can specify what SA will 'consider' a URI, or 
  to be able to 'capture' a value from an obfuscation test, manipulate it 
  into its 'original' URI and then 'manually' submit it to the URIBL

I did leak the other day, that I actually am hacking such a beast.

It works, but there's still some things to re-write properly. Stay
tuned. I'll announce it, when it is reasonably safe to use. Just be a
little bit patient, will ya? ;)

I was brief about this topic before, and I won't mention any details
today either. The above should be clear enough.

  guenther


-- 
char *t=\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4;
main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;il;i++){ i%8? c=1:
(c=*++x); c128  (s+=h); if (!(h=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-21 Thread Charles Gregory

Sometimes I wished everyone getting involved in heated discussions and
proposals, also would carefully read any post with a related topic...
I did leak the other day, that I actually am hacking such a beast.


Sorry. Sometimes the mailbox overload is a bit much, and I just have to 
delete things which 'seem' outside the central topics I'm following.

Still very glad to hear that something is in the works... :)


It works, but there's still some things to re-write properly. Stay
tuned. I'll announce it, when it is reasonably safe to use. Just be a
little bit patient, will ya? ;)


(smile) Thanks.

- charles


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-16 Thread Charles Gregory

On Wed, 15 Jul 2009, MrGibbage wrote:

I wonder if the spammers are reading this forum.  That seemed awful fast.


I'm sure they do. But I also suspect that they have a simple 'feedback' 
mechanism that let's them know how much of their spew is getting rejected
on their botnets, and when the rejection numbers get too high they try 
something new, and keep trying until the rejection numbers drop again.


Then we fix our rules, the rejections go up, and they look for yet another 
'trick' to get through. They have the advantage of being able to download 
their own copies of spamassassin to 'test' their spew. That's why 
sometimes you get 'red herrings' from me on this list when I don't share 
the full details of a rule. Posting it here almost assures that it will 
get bypassed. They copy the rule, then try all sorts of different 
combinations to bypass it


Now really, the significant factor here is not that any of these 
obfuscation tricks are 'new', but that they are using them to bypass the 
URIBL rules. I strongly urge the spamassassin develpopers to consider ways 
to 'open up' the way that we can specify what SA will 'consider' a URI, or 
to be able to 'capture' a value from an obfuscation test, manipulate it 
into its 'original' URI and then 'manually' submit it to the URIBL


Example hypothetical syntax (note that some parentheses are *capturing*):

body FINDURI /(www)(?:obfuscation)(domain)(?:obfuscation)(com|net|org)/i
uribl CHECIT /$1.$2.$3/

Basically, allow a rule to 'capture' one or more 'matches' in Perl 
variables, and then feed them to a subsequent rule (in this case, a manual 
URIBL lookup). This way, the SA developers don't have to hard-code an 
ever-changing set of URI detection rules into the core code, but we can 
still develop on-the-fly rules that can feed a URI to the URIBL tests


I've heard people mention 'plugins'. Could I code one that would be
easily 'modifiable' so that (for example) this morning's '[dot]' trick can 
be quickly added to my plugin? Is there a good working example of a plugin 
that extracts text from a message and feeds it to a URI? I'll work on 
this!


- C



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-15 Thread John Hardin

On Wed, 15 Jul 2009, MrGibbage wrote:


I wonder if the spammers are reading this forum.  That seemed awful fast.


Of course they are.

--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  USMC Rules of Gunfighting #20: The faster you finish the fight,
  the less shot you will get.
---
 Tomorrow: the 64th anniversary of the dawn of the Atomic Age


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-14 Thread Hrothgar

Which of course means we've long since passed the point where any of  
these are going to do the spammers any good.  That's the frustrating  
part.

I thought that the point was that since it cost a spammer the same to send
out a million emails as to send out one, he was happy if only one of the
recipients responded. 

I live in the UK. The chances of anyone here buying prescription drugs from
a web site are non-existent: they are paid for either by the health service
or (for those who have medical insurance) by insurers. And the, er, get it
up medicines are now available over the counter. Yet all co.uk addresses
get mountains of this type of spam which presumably sell nothing.

I find it quicker to delete them manually rather than spending time altering
a regex and restarting SA.

Roger
-- 
View this message in context: 
http://www.nabble.com/-NEW-SPAM-FLOOD--www.shopXX.net-tp24139422p24486959.html
Sent from the SpamAssassin - Users mailing list archive at Nabble.com.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread Charles Gregory


If I might interject. This seems to be an excellent occasion for
the PerlRE 'negative look-ahead' code (excuse the line wrap):

body =~ /(?!www\.[a-z]{2,6}[0-9]{2,6}\.(com|net|org))
www[^a-z0-9]+[a-z]{2,6}[0-9]{2,6}[^a-z0-9]+(com|net|org)/i

...unless someone can think of an FP for this rule?

- C


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread rich...@buzzhost.co.uk
On Mon, 2009-07-13 at 10:46 -0400, Charles Gregory wrote:
 (?!www\.[a-z]{2,6}[0-9]{2,6}\.(com|net|org))
 www[^a-z0-9]+[a-z]{2,6}[0-9]{2,6}[^a-z0-9]+(com|net|org)

Does not seem to work with;

www. meds .com



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread McDonald, Dan
On Mon, 2009-07-13 at 16:03 +0100, rich...@buzzhost.co.uk wrote:
 On Mon, 2009-07-13 at 10:46 -0400, Charles Gregory wrote:
  (?!www\.[a-z]{2,6}[0-9]{2,6}\.(com|net|org))
  www[^a-z0-9]+[a-z]{2,6}[0-9]{2,6}[^a-z0-9]+(com|net|org)
 
 Does not seem to work with;
 
 www. meds .com

It shouldn't.  The spammers have been using domains with 2-4 alpha
characters and 2 digits.

 
-- 
Daniel J McDonald, CCIE # 2495, CISSP # 78281, CNX
www.austinenergy.com


signature.asc
Description: This is a digitally signed message part


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread Charles Gregory

On Mon, 13 Jul 2009, rich...@buzzhost.co.uk wrote:

On Mon, 2009-07-13 at 10:46 -0400, Charles Gregory wrote:

(?!www\.[a-z]{2,6}[0-9]{2,6}\.(com|net|org))
www[^a-z0-9]+[a-z]{2,6}[0-9]{2,6}[^a-z0-9]+(com|net|org)


Does not seem to work with;
www. meds .com


Correct. With spaces being one of the possible obfuscation characters,
this otherwise 'broad' rule is limited to the cookie-cutter URL's with 
numeric suffixes in the hostnames - something unlikely to appear in 
conversational text like whether the [www can com]municate ideas... :)


- Charles




Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread John Hardin

On Mon, 13 Jul 2009, McDonald, Dan wrote:


On Mon, 2009-07-13 at 16:03 +0100, rich...@buzzhost.co.uk wrote:

On Mon, 2009-07-13 at 10:46 -0400, Charles Gregory wrote:

(?!www\.[a-z]{2,6}[0-9]{2,6}\.(com|net|org))
www[^a-z0-9]+[a-z]{2,6}[0-9]{2,6}[^a-z0-9]+(com|net|org)


Does not seem to work with;

www. meds .com


It shouldn't.  The spammers have been using domains with 2-4 alpha
characters and 2 digits.


Why be restrictive on the domain name?

\b(?!www\.\w{2,20}\.(?:com|net|org))www[^a-z0-9]+\w{2,20}[^a-z0-9]+(?:com|net|org)\b

The + signs are a little risky, it might be better to use {1,3} instead. 
And the older rule allowed for spaces in the TLD. I don't recall if 
anybody provided more than one spample with that though.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Users mistake widespread adoption of Microsoft Office for the
  development of a document format standard.
---
 3 days until the 64th anniversary of the dawn of the Atomic Age


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread John Hardin

On Mon, 13 Jul 2009, Charles Gregory wrote:


On Mon, 13 Jul 2009, rich...@buzzhost.co.uk wrote:

 On Mon, 2009-07-13 at 10:46 -0400, Charles Gregory wrote:
  (?!www\.[a-z]{2,6}[0-9]{2,6}\.(com|net|org))
  www[^a-z0-9]+[a-z]{2,6}[0-9]{2,6}[^a-z0-9]+(com|net|org)

 Does not seem to work with;
 www. meds .com


Correct. With spaces being one of the possible obfuscation characters,
this otherwise 'broad' rule is limited to the cookie-cutter URL's with 
numeric suffixes in the hostnames - something unlikely to appear in 
conversational text like whether the [www can com]municate ideas... :)


That possible FP is why \b are important in the rule.

--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Users mistake widespread adoption of Microsoft Office for the
  development of a document format standard.
---
 3 days until the 64th anniversary of the dawn of the Atomic Age


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread Charles Gregory

On Mon, 13 Jul 2009, John Hardin wrote:

Why be restrictive on the domain name?


If  a conservative spec is sufficient to match the spam, then we're
helping avoid false positives I'd rather tweak the rule to
catch the new tricks of the spammer than overgeneralize. :)

The + signs are a little risky, it might be better to use {1,3} instead.

(nod) Though without the '/m' option it would be limited to the same line.
My thinking is that a spammer would quickly figure out to add more 
obfuscation, and there is little risk of a false positive occuring with

that kind of broad spacing and an xxx99 domain name

And the older rule allowed for spaces in the TLD. I don't recall if 
anybody provided more than one spample with that though.


I've not seen it too much, though it doesn't hurt to keep it in the
rule. I actually added it back into my live rule after I posted

To answer your next post, I don't use '\b' because the next 'trick' coming 
will likely be something looking like Xwww herenn comX...  :)


- C


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread John Hardin

On Mon, 13 Jul 2009, Charles Gregory wrote:


On Mon, 13 Jul 2009, John Hardin wrote:

 Why be restrictive on the domain name?


If a conservative spec is sufficient to match the spam, then we're
helping avoid false positives I'd rather tweak the rule to
catch the new tricks of the spammer than overgeneralize. :)


Fair enough.

The + signs are a little risky, it might be better to use {1,3} 
instead.


(nod) Though without the '/m' option it would be limited to the same 
line.


body rules work on paragraphs, but you are right, the badness has an upper 
limit.


My thinking is that a spammer would quickly figure out to add more 
obfuscation, and there is little risk of a false positive occuring with 
that kind of broad spacing and an xxx99 domain name


Again, fair enough. But there's a limit to how complex the obfuscation can 
be made, though, because there's a point where people won't deobfuscate 
the URI to visit it.


To answer your next post, I don't use '\b' because the next 'trick' 
coming will likely be something looking like Xwww herenn comX...  :)


At that point it can be dealt with. Until then, using \b is an important 
way to avoid FPs.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Ignorance doesn't make stuff not exist.   -- Bucky Katt
---
 3 days until the 64th anniversary of the dawn of the Atomic Age


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread Charles Gregory

On Mon, 13 Jul 2009, John Hardin wrote:

  The + signs are a little risky, it might be better to use {1,3} instead.
 (nod) Though without the '/m' option it would be limited to the same line.
body rules work on paragraphs, but you are right, the badness has an upper 
limit.


Ugh. Forgot it was 'paragraphs' and not 'lines' (and I just had that 
drilled into me recently, too). Paragraphs are too long. I'll switch it

to a specific limit


 To answer your next post, I don't use '\b' because the next 'trick' coming
 will likely be something looking like Xwww herenn comX...  :)

At that point it can be dealt with.


Well, they're getting close. I'm seeing non-alpha non-blank crud cozied up 
to the front of the 'www' now :)


- C


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-13 Thread Cedric Knight
Chris Owen wrote:
 On Jul 13, 2009, at 2:55 PM, Charles Gregory wrote:
 
 To answer your next post, I don't use '\b' because the next 'trick'
 coming
 will likely be something looking like Xwww herenn comX...  :)
 At that point it can be dealt with.
 
 Well, they're getting close. I'm seeing non-alpha non-blank crud
 cozied up to the front of the 'www' now :)

Not forgetting underscores are not word boundaries.  My alternative
rules are badly written but are still hitting with the \b:

rawbody NONLINK_SHORT
/^.{0,500}\b(?:H\s*T\s*T\s*P\s*[:;](?!http:)\W{0,10}|W\s{0,10}W\s{0,10}W\s{0,10}(?:[.,\'`_+\-]\s{0,10})?(?!www\.))[a-z0-9\-]{3,13}\s{0,10}(?:[.,\'`_+\-]\s{0,10})?(?![a-z0-9]\.)(?:net|c\s{0,10}o\s{0,10}m|org|info|biz)\b/si
describe NONLINK_SHORT  Obfuscated link near top of text
score NONLINK_SHORT 2.5

#quite strict:
rawbody NONLINK_VSHORT  /^.{0,100}\bwww{0,2}(?:\. | \.|
?[,*_\-\+] ?)[a-z]{2,5}[0-9\-]{1,5}(?:\. | \.| ?[,*_\-\+]
?)(?:net|c\s{0,10}o\s{0,10}m|org|info|biz)(?:\. \S|\s*$)/s
describe NONLINK_VSHORT Specific obfuscated link form near top
of text
score NONLINK_VSHORT2.5

(These use rawbody with a caret to limit the area of matching to the
first few lines.)

So how about dropping the \b and using something looser like: 'w
?w(?!\.[a-z0-9\-]{2,12}\.(?:com|info|net|org|biz))[[:punct:]X
]{1,4}[a-z0-9\-]{2,12}[[:punct:]X ]{1,4}(?:c ?o ?m|info|n ?e ?t|o ?r
?g|biz)([[:punct:]X ]|$)'   ...?

 
 
 Which of course means we've long since passed the point where any of
 these are going to do the spammers any good.  That's the frustrating part.

You're making the common assumption that spammers send UCE because it
makes them money.  In fact they do it because they are obnoxious
imbeciles who want to annoy people and waste as much time (human and
CPU) as possible.  I don't think it really matters to them that what
they are sending is incomprehensible noise, because noise is their message.

Cheers

CK


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-12 Thread Sim
2009/7/11 Sim simvi...@gmail.com:
 New rules:
 body    __MED_BEG_SP    /\bw{2,3}[[:space:]][[:alpha:]]{2,6}\d{2,6}/i
 body    __MED_BEG_PUNCT /\bw{2,3}[[:punct:]]{1,3}[[:alpha:]]{2,6}\d{2,6}/i
 body    __MED_BEG_DOT   /\bw{2,3}\.[[:alpha:]]{2,6}\d{2,6}/i
 body    __MED_BEG_BOTH
 /\bw{2,3}[[:punct:][:space:]]{2,5}[[:alpha:]]{2,6}\d{2,6}\b/i
 body    __MED_END_SP
 /[[:alpha:]]{2,6}\d{2,6}[[:space:]](?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 body    __MED_END_PUNCT
 /[[:alpha:]]{2,6}\d{2,6}[[:punct:]]{1,3}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 body    __MED_END_DOT
 /[[:alpha:]]{2,6}\d{2,6}\.(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 body    __MED_END_BOTH
 /[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,5}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i

 meta    AE_MED42    (__MED_BEG_SP || __MED_BEG_PUNCT || __MED_BEG_DOT ||
 __MED_BEG_BOTH )  (__MED_END_SP || __MED_END_PUNCT || __MED_END_DOT ||
 __MED_END_BOTH)  ! (__MED_BEG_DOT  __MED_END_DOT )
 describe AE_MED42   rule to catch still more spam obfuscation
 score   AE_MED42    4.0





 Hi Dan,
 very very thanks!
 It's perfect for all variant!

 Regards



Hi!

Again tipology.  Rpace Against the Clocck.www_ze44_com

:-(

Spammer observe us!


Re: [NEW SPAM FLOOD] www_nu26_com

2009-07-12 Thread Charles Gregory

On Sat, 11 Jul 2009, Jason L Tibbitts III wrote:

I still wonder, though, if we shouldn't be turning these back into
hostnames and looking them up in the regular URI blacklists


Given the obvious objections to having the primary URIBL mechanism try to 
parse obfuscations, I once again question why we cannot have some sort of 
mechanism for 'capturing' the values of ordinary tests (such as the overly 
comnplex rule to catch these uribl obfuscations) and then have that value 
to manually feed to another test? There would be some interesting details 
to such a thing, for instance, if a rule matches more than one obfuscated 
URI, the 'capture' mechansim would have to somehow 'deliver' each captured 
value as an iteration of any check/test that included it


But for cases like this URI stuff, something 'flexible' is needed

- Charles


RE: [NEW SPAM FLOOD] www.shopXX.net

2009-07-12 Thread John Hardin

On Fri, 10 Jul 2009, McDonald, Dan wrote:


They have.  They are using underscores, which are a [:punct:], but don't form a 
\b break.

New rules:
body__MED_BEG_SP/\bw{2,3}[[:space:]][[:alpha:]]{2,6}\d{2,6}/i
body__MED_BEG_PUNCT /\bw{2,3}[[:punct:]]{1,3}[[:alpha:]]{2,6}\d{2,6}/i
body__MED_BEG_DOT   /\bw{2,3}\.[[:alpha:]]{2,6}\d{2,6}/i
body__MED_BEG_BOTH  
/\bw{2,3}[[:punct:][:space:]]{2,5}[[:alpha:]]{2,6}\d{2,6}\b/i
body__MED_END_SP
/[[:alpha:]]{2,6}\d{2,6}[[:space:]](?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
body__MED_END_PUNCT 
/[[:alpha:]]{2,6}\d{2,6}[[:punct:]]{1,3}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
body__MED_END_DOT   
/[[:alpha:]]{2,6}\d{2,6}\.(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
body__MED_END_BOTH  
/[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,5}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i

metaAE_MED42(__MED_BEG_SP || __MED_BEG_PUNCT || __MED_BEG_DOT || __MED_BEG_BOTH ) 
 (__MED_END_SP || __MED_END_PUNCT || __MED_END_DOT || __MED_END_BOTH)  ! 
(__MED_BEG_DOT  __MED_END_DOT )
describe AE_MED42   rule to catch still more spam obfuscation
score   AE_MED424.0


I think that can be simplified somewhat by reversing the obfuscation 
matches:


body  URI_OBFU_WWW   
/\bw{2,3}[^[:alnum:]]{1,3}\w{1,20}(?:(?!\.[[:alnum:]])[^[:alnum:]]{1,3})(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
describe  URI_OBFU_WWW   Obfuscated URI


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  The world has enough Mouse Clicking System Engineers.
   -- Dave Pooser
---
 4 days until the 64th anniversary of the dawn of the Atomic Age


RE: [NEW SPAM FLOOD] www.shopXX.net

2009-07-11 Thread rich...@buzzhost.co.uk
On Fri, 2009-07-10 at 22:46 -0500, McDonald, Dan wrote:
 From: Jason L Tibbitts III [mailto:ti...@math.uh.edu]
  MD == McDonald, Dan dan.mcdon...@austinenergy.com writes:
 
 MD They are using underscores, which are a [:punct:], but don't form
 MD a \b break.
 
 I'm becoming confused as to what they could possibly hope to
 accomplish by that.
 
 right now I think they are sticking it to us.  That and they must get
 some
 sort of jollies describing sick sex acts to little old ladies.
 
 Yes, I know, don't question the motives of spammers for their
 stupidity and madness may be contagious, but still.  Surely they must
 expect some kind of click rate.
 
 I expect they will tire quickly of this game.  I was expecting commas
 before underscores, but even that is a loss now.  So, they will have
 to
 play a new game, and we can start all over with the fun.
 
 
 
One of my customers has this in their Postfix body blocks and it seems
to do well. No doubt it could be adapted to SA or even made more 'curt'

/www((\.\s{1,10}|\s{1,10}\.|
\s{1,10}\.\s{1,10})[a-z1-9]{1,50}(\.\s{1,10}|\s{1,10}\.|
\s{1,10}\.\s{1,10}|\.)|\.[a-z1-9]{1,50}(\.\s{1,10}|\s{1,10}\.|
\s{1,10}\.\s{1,10}))(net|com)/REJECT body contains officated uri

Use it at your own risk



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-11 Thread Sim
 New rules:
 body    __MED_BEG_SP    /\bw{2,3}[[:space:]][[:alpha:]]{2,6}\d{2,6}/i
 body    __MED_BEG_PUNCT /\bw{2,3}[[:punct:]]{1,3}[[:alpha:]]{2,6}\d{2,6}/i
 body    __MED_BEG_DOT   /\bw{2,3}\.[[:alpha:]]{2,6}\d{2,6}/i
 body    __MED_BEG_BOTH
 /\bw{2,3}[[:punct:][:space:]]{2,5}[[:alpha:]]{2,6}\d{2,6}\b/i
 body    __MED_END_SP
 /[[:alpha:]]{2,6}\d{2,6}[[:space:]](?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 body    __MED_END_PUNCT
 /[[:alpha:]]{2,6}\d{2,6}[[:punct:]]{1,3}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 body    __MED_END_DOT
 /[[:alpha:]]{2,6}\d{2,6}\.(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 body    __MED_END_BOTH
 /[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,5}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i

 meta    AE_MED42    (__MED_BEG_SP || __MED_BEG_PUNCT || __MED_BEG_DOT ||
 __MED_BEG_BOTH )  (__MED_END_SP || __MED_END_PUNCT || __MED_END_DOT ||
 __MED_END_BOTH)  ! (__MED_BEG_DOT  __MED_END_DOT )
 describe AE_MED42   rule to catch still more spam obfuscation
 score   AE_MED42    4.0





Hi Dan,
very very thanks!
It's perfect for all variant!

Regards

---
Sim


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-11 Thread Paweł Tęcza
Dnia 2009-07-10, pią o godzinie 16:48 -0700, fchan pisze:
 Don't tempt them, I already get enough spam not only from these guys.
 Also they will flood the network with smtp useless connections and
 unless you have good network attack mitigation system so you don't
 have a DDoS, don't tempt them.

Please don't be affraid and help to beat them.

Do you only update your local rules? I think it's not sufficient
reaction. We also should send abuse reports to Internet providers of
spammers. They have to shutdown that website.

P.




RE: [NEW SPAM FLOOD] www.shopXX.net

2009-07-11 Thread McDonald, Dan
From: rich...@buzzhost.co.uk [mailto:rich...@buzzhost.co.uk]
On Fri, 2009-07-10 at 22:46 -0500, McDonald, Dan wrote:
 From: Jason L Tibbitts III [mailto:ti...@math.uh.edu]
  MD == McDonald, Dan dan.mcdon...@austinenergy.com writes:
 
 MD They are using underscores, which are a [:punct:], but don't form
 MD a \b break.

One of my customers has this in their Postfix body blocks and it seems
to do well. No doubt it could be adapted to SA or even made more 'curt'

/www((\.\s{1,10}|\s{1,10}\.|
\s{1,10}\.\s{1,10})[a-z1-9]{1,50}(\.\s{1,10}|\s{1,10}\.|
\s{1,10}\.\s{1,10}|\.)|\.[a-z1-9]{1,50}(\.\s{1,10}|\s{1,10}\.|
\s{1,10}\.\s{1,10}))(net|com)/REJECT body contains officated uri

Use it at your own risk

it won't hit anything now.  They aren't using periods any more.  They 
switched to underscores last night, and commas this morning.  Be ready 
for exclamation points later today!  Their click rate has to be dropping
like a rock and the only purpose at this point is to annoy us.






RE: [NEW SPAM FLOOD] www.shopXX.net

2009-07-11 Thread rich...@buzzhost.co.uk
On Sat, 2009-07-11 at 07:14 -0500, McDonald, Dan wrote:
 From: rich...@buzzhost.co.uk [mailto:rich...@buzzhost.co.uk]
 On Fri, 2009-07-10 at 22:46 -0500, McDonald, Dan wrote:
  From: Jason L Tibbitts III [mailto:ti...@math.uh.edu]
   MD == McDonald, Dan dan.mcdon...@austinenergy.com writes:
 
  MD They are using underscores, which are a [:punct:], but don't
 form
  MD a \b break.
 
 One of my customers has this in their Postfix body blocks and it
 seems
 to do well. No doubt it could be adapted to SA or even made more
 'curt'
 
 /www((\.\s{1,10}|\s{1,10}\.|
 \s{1,10}\.\s{1,10})[a-z1-9]{1,50}(\.\s{1,10}|\s{1,10}\.|
 \s{1,10}\.\s{1,10}|\.)|\.[a-z1-9]{1,50}(\.\s{1,10}|\s{1,10}\.|
 \s{1,10}\.\s{1,10}))(net|com)/REJECT body contains officated
 uri
 
 Use it at your own risk
 
 it won't hit anything now.  They aren't using periods any more.  They
 switched to underscores last night, and commas this morning.  Be ready
 for exclamation points later today!  Their click rate has to be
 dropping
 like a rock and the only purpose at this point is to annoy us.
 
I guess it goes without saying to duplicate the rule for other options ?
I've added duplicates for all the obvious characters on the keyboard -
I'm just waiting to see some more creativity from them :-)
 
 
 
 



Re: [NEW SPAM FLOOD] www_nu26_com

2009-07-11 Thread Jason L Tibbitts III
 MD == McDonald, Dan dan.mcdon...@austinenergy.com writes:

MD The rules I posted last night catch those.  They switched from
MD underscores to commas this morning, and my rules still catch them.

FYI, they're also using plus signs, which also seem to be caught
properly by your rules.  I think we're good until they switch to
alphanumerics like wwwZnu26Ycom, which we should be able to filter out
pretty trivially.

I still wonder, though, if we shouldn't be turning these back into
hostnames and looking them up in the regular URI blacklists, because
the looser we make the rules, the larger the chance of false
positives.  Not sure if spamassassin actually permits that, however.

 - J


RE: [NEW SPAM FLOOD] www_nu26_com

2009-07-11 Thread McDonald, Dan
From: Jason L Tibbitts III [mailto:ti...@math.uh.edu]
 MD == McDonald, Dan dan.mcdon...@austinenergy.com writes:

MD The rules I posted last night catch those.  They switched from
MD underscores to commas this morning, and my rules still catch them.

I still wonder, though, if we shouldn't be turning these back into
hostnames and looking them up in the regular URI blacklists, because
the looser we make the rules, the larger the chance of false
positives. 

That's why I have the exclude two dots part of the rule.  My first attempt 
was getting a lot of false positives.  Anyone obfuscating the domain name, 
IMHO, is definitely asking to be blocked.

--
Dan McDonald, CCIE # 2495, CISSP # 78281, CNX



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Sim


 /\bwww(?:\s|\s\W|\W\s)\w{3,6}\d{2,6}(?:\s|s\W|\W\s)(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i

                                           ^
 John,

 Thanks a lot for rule update! It works fine. I can say it's nearly
 perfect, because it missing only one small back-slash :) Please look
 above.

 D'oh!

 That, plus some other fixes:

 /\bwww(?:\s\W?\s?|\W\s)\w{3,6}\d{2,6}(?:\s\W?\s?|\W\s)(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i



Hello world ;-)

I'm using it without good results for this format:

bla bla www. site. net. bla bla

Have you any idea?
Regards

---
Sim


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread McDonald, Dan
On Fri, 2009-07-10 at 17:11 +0200, Sim wrote:
 
 
  /\bwww(?:\s|\s\W|\W\s)\w{3,6}\d{2,6}(?:\s|s\W|\W\s)(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 
^
  John,
 
  Thanks a lot for rule update! It works fine. I can say it's nearly
  perfect, because it missing only one small back-slash :) Please look
  above.
 
  D'oh!
 
  That, plus some other fixes:
 
  /\bwww(?:\s\W?\s?|\W\s)\w{3,6}\d{2,6}(?:\s\W?\s?|\W\s)(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 
 
 
 Hello world ;-)
 
 I'm using it without good results for this format:
 
 bla bla www. site. net. bla bla
 
 Have you any idea?
 Regards
Yes, remove the outer parentheses.

Here are the rules I am using:
bodyAE_MEDS35   /w{2,4}\s(?:meds|shop)\d{1,4}\s(?:net|com|org)/
describe AE_MEDS35  obfuscated domain seen in spam
score   AE_MEDS35   3.00

bodyAE_MEDS38   
/\(\s?w{2,4}\s[[:alpha:]]{4}\d{1,4}\s(?:net|com|org)\s?\)/
describe AE_MEDS38  rule to catch next wave of obfuscated domains
score   AE_MEDS38   1.0

bodyAE_MEDS39   
/\bw{2,3}[[:punct:][:space:]]{2,3}[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,3}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
describe AE_MEDS39  rule to catch still more spam obfuscation
score   AE_MEDS39   4.0

AE_MEDS38 finds domains with spaces in them, and AE_MEDS39 finds domains
with dots and spaces.  You might want to bump up the score on AE_MEDS38,
but I haven't had a false negative that would have benefited from it in
a while, so I haven't bothered.



-- 
Daniel J McDonald, CCIE # 2495, CISSP # 78281, CNX
www.austinenergy.com


signature.asc
Description: This is a digitally signed message part


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Daniel Schaefer

McDonald, Dan wrote:

Yes, remove the outer parentheses.

Here are the rules I am using:
bodyAE_MEDS35   /w{2,4}\s(?:meds|shop)\d{1,4}\s(?:net|com|org)/
describe AE_MEDS35  obfuscated domain seen in spam
score   AE_MEDS35   3.00

bodyAE_MEDS38   
/\(\s?w{2,4}\s[[:alpha:]]{4}\d{1,4}\s(?:net|com|org)\s?\)/
describe AE_MEDS38  rule to catch next wave of obfuscated domains
score   AE_MEDS38   1.0

bodyAE_MEDS39   
/\bw{2,3}[[:punct:][:space:]]{2,3}[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,3}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
describe AE_MEDS39  rule to catch still more spam obfuscation
score   AE_MEDS39   4.0

  

Since we're sharing rules for this recent Spam outbreak, here is my rule:
body DRUG_SITE /www(\.|\ 
)*(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\ )*(net|com)/

score DRUG_SITE 0.5
describe DRUG_SITE Test to find spam drug sites in recent emails


Notice my score is low, because I'm not sure it's 100% accurate.

--
Dan Schaefer
Application Developer
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread John Hardin

On Fri, 10 Jul 2009, Sim wrote:


/\bwww(?:\s\W?\s?|\W\s)\w{3,6}\d{2,6}(?:\s\W?\s?|\W\s)(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i


I'm using it without good results for this format:

bla bla www. site. net. bla bla

Have you any idea?


There are no digits in that URI.

If this becomes common, change the \d{2,6} to \d{0,6}, but that will 
increase the risk of FP somewhat.


Dan: there are no parentheses in that RE that attempt to match the message 
text, they are all grouping parentheses.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  The Constitution is a written instrument. As such its meaning does
  not alter. That which it meant when adopted, it means now.
-- U.S. Supreme Court
   SOUTH CAROLINA v. US, 199 U.S. 437, 448 (1905)
---
 10 days until the 40th anniversary of Apollo 11 landing on the Moon


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread McDonald, Dan
On Fri, 2009-07-10 at 11:39 -0400, Daniel Schaefer wrote:
 McDonald, Dan wrote:

 Since we're sharing rules for this recent Spam outbreak, here is my rule:
 body DRUG_SITE /www(\.|\ 
 )*(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\ )*(net|com)/

You should avoid the use of *, as it allows spammers to consume all of
your memory and cpu.  limit it using the {} syntax.  You also should
tell perl to not keep the results of your () with (?:\.|\ ) instead of
(\.|\ ).  And with single characters, the [ab] syntax is faster to
process than (?:a|b).




-- 
Daniel J McDonald, CCIE # 2495, CISSP # 78281, CNX
www.austinenergy.com


signature.asc
Description: This is a digitally signed message part


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Gerry Maddock
  McDonald, Dan wrote:

  Since we're sharing rules for this recent Spam outbreak, here is my
rule:
  body DRUG_SITE /www(\.|\
  )*(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\ )*(net|
com)/

 You should avoid the use of *, as it allows spammers to consume all of
 your memory and cpu.  limit it using the {} syntax.  You also should
 tell perl to not keep the results of your () with (?:\.|\ ) instead of
 (\.|\ ).  And with single characters, the [ab] syntax is faster to
 process than (?:a|b).

Perhaps you could attach an example showing exactly what your stating for
this rule?





CONFIDENTIALITY: This e-mail message is for the sole use of the intended 
recipient(s) and may contain confidential and / or privileged information.  Any 
unauthorized review, use, disclosure or distribution of any kind is strictly 
prohibited.  If you are not the intended recipient, please contact the sender 
via reply e-mail and destroy all copies of the original message.  Thank you.






Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Daniel Schaefer

Gerry Maddock wrote:

McDonald, Dan wrote:
  
Since we're sharing rules for this recent Spam outbreak, here is my
  

rule:
  

body DRUG_SITE /www(\.|\
)*(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\ )*(net|
  

com)/
  

You should avoid the use of *, as it allows spammers to consume all of
your memory and cpu.  limit it using the {} syntax.  You also should
tell perl to not keep the results of your () with (?:\.|\ ) instead of
(\.|\ ).  And with single characters, the [ab] syntax is faster to
process than (?:a|b).



Perhaps you could attach an example showing exactly what your stating for
this rule?

  

This is my new rule. I think this is what he means:

body DRUG_SITE /www[\.\ 
]*(?:med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}[\.\ 
]*(?:net|com)/


--
Dan Schaefer
Application Developer
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Sim
 Yes, remove the outer parentheses.

 Here are the rules I am using:
 body    AE_MEDS35       /w{2,4}\s(?:meds|shop)\d{1,4}\s(?:net|com|org)/
 describe AE_MEDS35      obfuscated domain seen in spam
 score   AE_MEDS35       3.00

 body    AE_MEDS38       
 /\(\s?w{2,4}\s[[:alpha:]]{4}\d{1,4}\s(?:net|com|org)\s?\)/
 describe AE_MEDS38      rule to catch next wave of obfuscated domains
 score   AE_MEDS38       1.0

 body    AE_MEDS39       
 /\bw{2,3}[[:punct:][:space:]]{2,3}[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,3}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i
 describe AE_MEDS39      rule to catch still more spam obfuscation
 score   AE_MEDS39       4.0

 AE_MEDS38 finds domains with spaces in them, and AE_MEDS39 finds domains
 with dots and spaces.  You might want to bump up the score on AE_MEDS38,
 but I haven't had a false negative that would have benefited from it in
 a while, so I haven't bothered.




Very good!
Thanks a lot!

Regards and good week-end!

---
Sim


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Sim
2009/7/10 John Hardin jhar...@impsec.org:
 On Fri, 10 Jul 2009, Sim wrote:


 /\bwww(?:\s\W?\s?|\W\s)\w{3,6}\d{2,6}(?:\s\W?\s?|\W\s)(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i

 I'm using it without good results for this format:

 bla bla www. site. net. bla bla

 Have you any idea?

 There are no digits in that URI.

 If this becomes common, change the \d{2,6} to \d{0,6}, but that will
 increase the risk of FP somewhat.

 Dan: there are no parentheses in that RE that attempt to match the message
 text, they are all grouping parentheses.



Good solution John,

very thanks!

Regards

---
Sim


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread John Hardin

On Fri, 10 Jul 2009, Daniel Schaefer wrote:


Gerry Maddock wrote:

   McDonald, Dan wrote:
 
   body DRUG_SITE /www(\.|\
  ) *(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\ 
  ) )*(net|com)/
 
  You should avoid the use of *, as it allows spammers to consume all 
  of your memory and cpu.  limit it using the {} syntax.  You also 
  should tell perl to not keep the results of your () with (?:\.|\ ) 
  instead of (\.|\ ).  And with single characters, the [ab] syntax is 
  faster to process than (?:a|b).


 Perhaps you could attach an example showing exactly what your stating
 for this rule?


This is my new rule. I think this is what he means:

body DRUG_SITE /www[\.\ 
] *(?:med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}[\.\ *(?:net|com)/


You missed some of the suggestions.

Try this:

body DRUG_SITE 
/\bwww[.\s]{1,3}(?:med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)\d{2}[.\s]{1,3}(?:net|com)\b/

Also, if the spammers start registering three-digit domain names, this 
will start missing. Something like \d{2,5} would be better.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Gun Control laws cannot reduce violent crime, because gun control
  laws focus obsessively on a tool a criminal might use to commit a
  crime rather than the criminal himself and his act of violence.
---
 10 days until the 40th anniversary of Apollo 11 landing on the Moon


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Daniel Schaefer

John Hardin wrote:

On Fri, 10 Jul 2009, Daniel Schaefer wrote:


Gerry Maddock wrote:

   McDonald, Dan wrote:
 
   body DRUG_SITE /www(\.|\
  ) *(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\  
 ) )*(net|com)/
   You should avoid the use of *, as it allows spammers to consume 
all   of your memory and cpu.  limit it using the {} syntax.  You 
also   should tell perl to not keep the results of your () with 
(?:\.|\ )   instead of (\.|\ ).  And with single characters, the 
[ab] syntax is   faster to process than (?:a|b).


 Perhaps you could attach an example showing exactly what your stating
 for this rule?


This is my new rule. I think this is what he means:

body DRUG_SITE /www[\.\ ] 
*(?:med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}[\.\ 
*(?:net|com)/


You missed some of the suggestions.

Try this:

body DRUG_SITE 
/\bwww[.\s]{1,3}(?:med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)\d{2}[.\s]{1,3}(?:net|com)\b/ 



Also, if the spammers start registering three-digit domain names, this 
will start missing. Something like \d{2,5} would be better.



Doesn't the . (period) need escaped in this? [.\s]{1,3}

--
Dan Schaefer
Application Developer
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread John Hardin

On Fri, 10 Jul 2009, Daniel Schaefer wrote:


Doesn't the . (period) need escaped in this? [.\s]{1,3}


Nope. [] means explicit set of characters, and . = any character 
conflicts with that context.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Gun Control laws cannot reduce violent crime, because gun control
  laws focus obsessively on a tool a criminal might use to commit a
  crime rather than the criminal himself and his act of violence.
---
 10 days until the 40th anniversary of Apollo 11 landing on the Moon


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Daniel Schaefer

John Hardin wrote:

On Fri, 10 Jul 2009, Daniel Schaefer wrote:


Doesn't the . (period) need escaped in this? [.\s]{1,3}


Nope. [] means explicit set of characters, and . = any 
character conflicts with that context.



Thanks for the clarification. I'm still learning REs.

--
Dan Schaefer
Application Developer
Performance Administration Corp.



Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread Michelle Konzack
Am 2009-07-10 11:39:02, schrieb Daniel Schaefer:
 Since we're sharing rules for this recent Spam outbreak, here is my rule:
 body DRUG_SITE /www(\.|\  
 )*(med|meds|gen|pill|shop|via|cu|co|ba|da|bu|ba)[0-9]{2}(\.|\ 
 )*(net|com)/
 score DRUG_SITE 0.5
 describe DRUG_SITE Test to find spam drug sites in recent emails


 Notice my score is low, because I'm not sure it's 100% accurate.

Does not hit:

Problems in Getting the sex Life Ymoou Want and Deserve - Starting With E 
www.ma29. net. Californian Finds Pit Blul Under hTe Hood

Thanks, Greetings and nice Day/Evening
Michelle Konzack
Systemadministrator
Tamay Dogan Network
Debian GNU/Linux Consultant

-- 
Linux-User #280138 with the Linux Counter, http://counter.li.org/
# Debian GNU/Linux Consultant #
http://www.tamay-dogan.net/ Michelle Konzack
http://www.can4linux.org/   c/o Vertriebsp. KabelBW
http://www.flexray4linux.org/   Blumenstrasse 2
Jabber linux4miche...@jabber.ccc.de   77694 Kehl/Germany
IRC #Debian (irc.icq.com) Tel. DE: +49 177 9351947
ICQ #328449886Tel. FR: +33  6  61925193


signature.pgp
Description: Digital signature


Re: [NEW SPAM FLOOD] www.shopXX.net

2009-07-10 Thread John Hardin

On Fri, 10 Jul 2009, McDonald, Dan wrote:


body__MED_END_BOTH  
/\b[[:alpha:]]{2,6}\d{2,6}[[:punct:][:space:]]{2,5}(?:c\s?o\s?m|n\s?e\s?t|o\s?r\s?g)\b/i

Let's see how long it takes them to come up with a workaround for this!


A domain name with 7+ letters? www. goodmeds123. com ?  :)

--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  If healthcare is a Right means that the government is obligated
  to provide the people with hospitals, physicians, treatments and
  medications at low or no cost, then the right to free speech means
  the government is obligated to provide the people with printing
  presses and public address systems, the right to freedom of
  religion means the government is obligated to build churches for the
  people, and the right to keep and bear arms means the government is
  obligated to provide the people with guns, all at low or no cost.
---
 10 days until the 40th anniversary of Apollo 11 landing on the Moon


  1   2   3   >