Re: replay RBL queries one hour later

2023-02-26 Thread Rob McEwen

Benny,

All I know for sure is this - for MANY legit emails - DKIM fails some 
days later - when it had originally worked/validated at the time the 
message was sent. I see this often in the real world when I rescan a 
message to try to verify the impact on a message that a spam filtering 
change caused - then notice that a very legit email that original passed 
DKIM at the time the message was received - now suddenly fails DKIM 
during this days-later rescan - and without ANY changes to the message 
itself. I think that this is most likely caused by DNS records for that 
DKIM being changed/updated. But whatever the cause, this is STILL a 
reality that's worth noting, for anyone who is rescanning messages 
later.


Rob McEwen, invaluement


-- Original Message --

From "Benny Pedersen" 

To users@spamassassin.apache.org
Date 2/26/2023 1:37:53 PM
Subject Re: replay RBL queries one hour later


Rob McEwen skrev den 2023-02-26 19:03:

..

sent. This can lead to many egregious false positives. But doing this
"one hour later" shouldn't have this problem.


message-id is timebased, so why invalidate it ? :)

i did that mistake on not dkim sign that header

in that regard i now have 2048 kbit size, where 4096 is a bit overkill


Re: replay RBL queries one hour later

2023-02-26 Thread Rob McEwen
Something to keep in mind about this idea of rescanning messages later - 
once more anti-spam data is available - for use in training/reporting 
spams - this probably should NOT be done days later because SOME senders 
aggressively expire/recycle DKIM dns records. I guess that is to 
minimize the ability for criminals to spoof DKIM? The result is that if 
you implement this idea on days-old messages, you can end up with some 
spam scoring that was ONLY due to the DKIM not being valid anymore, 
where it was valid at the time the message was sent. This can lead to 
many egregious false positives. But doing this "one hour later" 
shouldn't have this problem.


Rob McEwen, invaluement


Re: May I get to 0 phishing?

2023-02-21 Thread Rob McEwen
I dug a little deeper on this. I'm pretty sure that FROM_PAYPAL_SPOOF is 
triggered at least in part by __NOT_SPOOFED being set to "false" - and 
DKIM failing does (or can) cause __NOT_SPOOFED to be false - and so in 
this case a failed DKIM validation, that most likely worked when the 
message was originally sent - is what's now causing this chain reaction. 
It's highly doubtful that this rule would have hit at the time the 
message was received.

--Rob McEwen, invaluement



-- Original Message --

From "Rob McEwen" 

To users@spamassassin.apache.org
Date 2/21/2023 4:53:27 PM
Subject Re: May I get to 0 phishing?


Benny,

There are a few holes in your theory/assertions:

(1) I know for a fact that this came from PayPal's official transactional servers, in 
PayPal's IP space. And while the sender (PayPal's customer) was a "bad actor", 
this wasn't PayPal's actual email server getting hacked. Instead, it was PayPal's 
deliberate notification they sent on purpose, with all the proper authentication that 
normally is sent in ALL legit PayPal emails.

(2) I'm about 99.9% certain that all the validations that fail now - passed 
when it was originally sent/received. It's actually common for such large 
senders to expire DKIM record validation either quickly (to make spoofing 
harder!) and/or to manually expire it when they find fraud in recently-sent 
spams. One or the other, or both, likely happened here. I'm very confident that 
some (probably all!) of the validation failures that caused some portion of 
your bad scoring - weren't there if SA had been run against this soon after it 
was sent.

(3) I'm using SA 4.x, and a few minutes ago, I ran this against SA, and I ran a legit PayPal 
notification from this SAME IP address, that was sent today - both against SA. 
"FROM_PAYPAL_SPOOF" never had a hit on either one - but I also have RBLs and URI-lists 
set to not run in my SA, since I'm doing all that elsewhere - so maybe that disabled 
FROM_PAYPAL_SPOOF in my system? Or maybe FROM_PAYPAL_SPOOF isn't in SA4? Nevertheless, if this rule 
is so great and definitive, why is it only scoring 1.2 points? 1.2 points suggests that it might 
not be 100% immune to false positives! And if your argument is so great, why was your overall SA 
score ONLY 1.2 points? Do you really think that everyone using SA should "know" to 
magically block all messages that ONLY score 1.3 points, but have a hit on this rule? Should other 
SA users have this magical insight about other such SA rules?

I think you destroyed your own argument, with your own evidence. And you seem 
to be overlooking the fact that these are sent from PayPal servers that also 
send a MASSIVE amount of legit and transactional emails, including from this 
actual same IP. For example, in the past 24 hours, my small-ish mail hosting 
system has 6 legit not-spam PayPal notifications sent from this SAME ip address 
- all 6 of those were legit.

Rob McEwen, invaluement



-- Original Message --
From "Benny Pedersen" 
To users@spamassassin.apache.org
Date 2/21/2023 4:03:31 PM
Subject Re: May I get to 0 phishing?


Rob McEwen skrev den 2023-02-21 20:37:


https://pastebin.com/v80qMF99


Content preview:  Invoice from Apple. com (0005) xxx...@example.com, here are
   your invoice details Hello, xxx...@example.com Here's your invoice

Content analysis details:   (1.2 points, 5.0 required)

 pts rule name  description
 -- --
-0.0 RCVD_IN_MSPIKE_H2  RBL: Average reputation (+2)
[173.0.84.227 listed in wl.mailspike.net]
 1.8 DKIM_ADSP_DISCARD  No valid author signature, domain signs all mail
and suggests discarding the rest
 0.1 DKIM_SIGNEDMessage has a DKIM or DK signature, not necessarily 
valid
 0.1 DKIM_INVALID   DKIM or DK signature exists, but is not valid
-2.3 RCVD_IN_DNSWL_MED  RBL: Sender listed at https://www.dnswl.org/,
medium trust
[173.0.84.227 listed in list.dnswl.org]
-2.0 RCVD_IN_HOSTKARMA_WRBL: Sender listed in HOSTKARMA-WHITE
[173.0.84.227 listed in hostkarma.junkemailfiltercom]
 0.0 HTML_MESSAGE   BODY: HTML included in message
 0.1 MIME_HTML_ONLY BODY: Message only has text/html MIME parts
 0.4 KAM_REALLYHUGEIMGSRC   RAW: Spam with image tags with ridiculously huge
 http urls
 0.0 MIME_QP_LONG_LINE  RAW: Quoted-printable line longer than 76 chars
 0.0 LONG_IMG_URI   Image URI with very long path component - web bug?
 1.0 MSGID_NOFQDN2  Message-ID without a fully-qualified domain name
 0.5 LONGLINE   Line length exceeds 998 character limit, RFC 5322
 1.2 FROM_PAYPAL_SPOOF  From PayPal domain but matches SPOOFED
 0.2 KAM_HUGEIMGSRC Message co

Re: May I get to 0 phishing?

2023-02-21 Thread Rob McEwen

Benny,

There are a few holes in your theory/assertions:

(1) I know for a fact that this came from PayPal's official 
transactional servers, in PayPal's IP space. And while the sender 
(PayPal's customer) was a "bad actor", this wasn't PayPal's actual email 
server getting hacked. Instead, it was PayPal's deliberate notification 
they sent on purpose, with all the proper authentication that normally 
is sent in ALL legit PayPal emails.


(2) I'm about 99.9% certain that all the validations that fail now - 
passed when it was originally sent/received. It's actually common for 
such large senders to expire DKIM record validation either quickly (to 
make spoofing harder!) and/or to manually expire it when they find fraud 
in recently-sent spams. One or the other, or both, likely happened here. 
I'm very confident that some (probably all!) of the validation failures 
that caused some portion of your bad scoring - weren't there if SA had 
been run against this soon after it was sent.


(3) I'm using SA 4.x, and a few minutes ago, I ran this against SA, and 
I ran a legit PayPal notification from this SAME IP address, that was 
sent today - both against SA. "FROM_PAYPAL_SPOOF" never had a hit on 
either one - but I also have RBLs and URI-lists set to not run in my SA, 
since I'm doing all that elsewhere - so maybe that disabled 
FROM_PAYPAL_SPOOF in my system? Or maybe FROM_PAYPAL_SPOOF isn't in SA4? 
Nevertheless, if this rule is so great and definitive, why is it only 
scoring 1.2 points? 1.2 points suggests that it might not be 100% immune 
to false positives! And if your argument is so great, why was your 
overall SA score ONLY 1.2 points? Do you really think that everyone 
using SA should "know" to magically block all messages that ONLY score 
1.3 points, but have a hit on this rule? Should other SA users have this 
magical insight about other such SA rules?


I think you destroyed your own argument, with your own evidence. And you 
seem to be overlooking the fact that these are sent from PayPal servers 
that also send a MASSIVE amount of legit and transactional emails, 
including from this actual same IP. For example, in the past 24 hours, 
my small-ish mail hosting system has 6 legit not-spam PayPal 
notifications sent from this SAME ip address - all 6 of those were 
legit.


Rob McEwen, invaluement



-- Original Message --

From "Benny Pedersen" 

To users@spamassassin.apache.org
Date 2/21/2023 4:03:31 PM
Subject Re: May I get to 0 phishing?


Rob McEwen skrev den 2023-02-21 20:37:


https://pastebin.com/v80qMF99


Content preview:  Invoice from Apple. com (0005) xxx...@example.com, here are
   your invoice details Hello, xxx...@example.com Here's your invoice

Content analysis details:   (1.2 points, 5.0 required)

 pts rule name  description
 -- --
-0.0 RCVD_IN_MSPIKE_H2  RBL: Average reputation (+2)
[173.0.84.227 listed in wl.mailspike.net]
 1.8 DKIM_ADSP_DISCARD  No valid author signature, domain signs all mail
and suggests discarding the rest
 0.1 DKIM_SIGNEDMessage has a DKIM or DK signature, not necessarily 
valid
 0.1 DKIM_INVALID   DKIM or DK signature exists, but is not valid
-2.3 RCVD_IN_DNSWL_MED  RBL: Sender listed at https://www.dnswl.org/,
medium trust
[173.0.84.227 listed in list.dnswl.org]
-2.0 RCVD_IN_HOSTKARMA_WRBL: Sender listed in HOSTKARMA-WHITE
[173.0.84.227 listed in hostkarma.junkemailfilter.com]
 0.0 HTML_MESSAGE   BODY: HTML included in message
 0.1 MIME_HTML_ONLY BODY: Message only has text/html MIME parts
 0.4 KAM_REALLYHUGEIMGSRC   RAW: Spam with image tags with ridiculously huge
 http urls
 0.0 MIME_QP_LONG_LINE  RAW: Quoted-printable line longer than 76 chars
 0.0 LONG_IMG_URI   Image URI with very long path component - web bug?
 1.0 MSGID_NOFQDN2  Message-ID without a fully-qualified domain name
 0.5 LONGLINE   Line length exceeds 998 character limit, RFC 5322
 1.2 FROM_PAYPAL_SPOOF  From PayPal domain but matches SPOOFED
 0.2 KAM_HUGEIMGSRC Message contains many image tags with huge http urls
 0.1 DMARC_REJECT   DMARC reject policy
 0.0 LOTS_OF_MONEY  Huge... sums of money
 0.0 T_REMOTE_IMAGE Message contains an external image

dont know more, but dnswl ? ;)

DMARC_REJECT && FROM_PAYPAL_SPOOF why accept it ?



Re: May I get to 0 phishing?

2023-02-21 Thread Rob McEwen
Nope. That was a phishing spam, just maybe not the TYPE of phishing spam 
you're used to seeing? Calling it a fraud doesn't make it not a phish. 
When is a phishing spam ever NOT fraud? So what's the deciding factor? 
The fact that this claimed to be Apple sending an invoice via PayPal - 
and tried to trick the end user into thinking that they were the real 
Apple - except it wasn't really Apple - it was a criminal masquerading 
as Apple, trying to trick the recipient into paying this via thinking 
that this was the real Apple. THAT is the deciding factor that makes 
this a phish as well as a fraud.


(PayPal should have done better customer vetting on the front end!)

Rob McEwen, invaluement


-- Original Message --

From "hg user" 

To "Rob McEwen" 
Cc users@spamassassin.apache.org
Date 2/21/2023 3:10:35 PM
Subject Re: May I get to 0 phishing?

I think this is not a phishing, more a fraud: it seems a real invoice 
for something you didn't buy.


I'm glad to hear from experts that it's impossible to have 0 phishing, 
that I'm not missing the "silver bullet" or the magic token.


I may perhaps implement ESP plugin, and subscribe to DQS, or add a OCR 
plugin for those very annoying "pay the fine" scams. Really dubious 
about enabling razor and pyzor for italian language.


Unfortunately my spamassassin, version 3.4.5, is embedded into Zimbra, 
and it makes me really afraid of adding plugins...


Suggestions are always welcome

On Tue, Feb 21, 2023 at 8:37 PM Rob McEwen  wrote:

What Bill Cole said! Agreed. For example, here's an almost impossible
phish to block (at least, without blocking legitimate PayPal
transactional emails!). This is a PayPal phishing spam, sent from
PayPal's own server! It was sent by PayPal. I only changed the 
intended
recipient address (to protect the innocent), and changed the "=" at 
the

end of lines MIME-formatting to regular lines, for better readability
when looking through the email body for links. Otherwise, not altered.

https://pastebin.com/v80qMF99

However - there are always very helpful improvements that can be made
for minimizing the number of phish that get into the inbox. It's a
constant battle!

Rob McEwen, invaluement

-- Original Message --
From "Bill Cole" 
To users@spamassassin.apache.org
Date 2/21/2023 2:11:02 PM
Subject Re: May I get to 0 phishing?

>On 2023-02-21 at 13:51:09 UTC-0500 (Tue, 21 Feb 2023 19:51:09 +0100)
>hg user 
>is rumored to have said:
>
>>I was wondering if it is possible to reach the goal of 0 phishing.
>
>Nope. There are people who find it profitable and they will continue 
to find ways to trick all the usable programmatic mechanisms deployed 
to stop it.

>
>>With 2 layers of paid protection, and a third layer realized with
>>spamassassin with a lot of hand made rules, I'm able to catch a lot 
of spam

>>and if some reaches the mailboxes, no problem.
>>
>>But when phishing is able to reach the mailboxes, it is more 
dangerous, and

>>I'd like to bring it to a minimum.
>>
>>I'd like to know if you, despite all the barriers, still, although 
rarely,

>>have phishing go through, and how do you handle the situation.
>
>Eternal vigilance and user education.
>
>The world is an imperfect place.
>
>
>-- Bill Cole
>b...@scconsult.com or billc...@apache.org
>(AKA @grumpybozo and many *@billmail.scconsult.com addresses)
>Not Currently Available For Hire

Re: May I get to 0 phishing?

2023-02-21 Thread Rob McEwen
What Bill Cole said! Agreed. For example, here's an almost impossible 
phish to block (at least, without blocking legitimate PayPal 
transactional emails!). This is a PayPal phishing spam, sent from 
PayPal's own server! It was sent by PayPal. I only changed the intended 
recipient address (to protect the innocent), and changed the "=" at the 
end of lines MIME-formatting to regular lines, for better readability 
when looking through the email body for links. Otherwise, not altered.


https://pastebin.com/v80qMF99

However - there are always very helpful improvements that can be made 
for minimizing the number of phish that get into the inbox. It's a 
constant battle!


Rob McEwen, invaluement

-- Original Message --

From "Bill Cole" 

To users@spamassassin.apache.org
Date 2/21/2023 2:11:02 PM
Subject Re: May I get to 0 phishing?


On 2023-02-21 at 13:51:09 UTC-0500 (Tue, 21 Feb 2023 19:51:09 +0100)
hg user 
is rumored to have said:


I was wondering if it is possible to reach the goal of 0 phishing.


Nope. There are people who find it profitable and they will continue to find 
ways to trick all the usable programmatic mechanisms deployed to stop it.


With 2 layers of paid protection, and a third layer realized with
spamassassin with a lot of hand made rules, I'm able to catch a lot of spam
and if some reaches the mailboxes, no problem.

But when phishing is able to reach the mailboxes, it is more dangerous, and
I'd like to bring it to a minimum.

I'd like to know if you, despite all the barriers, still, although rarely,
have phishing go through, and how do you handle the situation.


Eternal vigilance and user education.

The world is an imperfect place.


-- Bill Cole
b...@scconsult.com or billc...@apache.org
(AKA @grumpybozo and many *@billmail.scconsult.com addresses)
Not Currently Available For Hire


Re[2]: URIDNSBL full message checking

2023-02-06 Thread Rob McEwen




It's actually just a domain name.  This uridnsbl keys off domain names
in the body too, I was kinda hoping it would look at the domain names
in the headers like the body, guess not.


So there's an interesting history here. Back in the early/mid 2000s, 
when SURBL, URIBL, and invaluement's URI lists were just starting (I was 
there!) - we didn't have reliable and universally-used/established 
domain authentication tools like SPF and DKIM and even ESPs were either 
non-existent or just beginning. Therefore, the vast majority of spammers 
were sending from their own servers (or bots!) - and both the mail 
header from and the SMTP-envelope FROM - in spams - was 99+% of the time 
forged. So trying to run a DNSBL that listed the domains found in the 
headers was a horrible idea because a massive percentage of spam used 
forged domains. That was then a losing game of whack-a-mole that would 
only add much useless one-off data to a dnsbl, as well as providing 
spammers with intel they could use to find DNSBL spamtrap addresses.


Today, so much is radically different since now many spams have their 
domains authenticated with things like SPF and DKIM. Therefore, SURBL 
and URIBL and Spamhaus's DBL have since moved more towards purposely 
including those header and SMTP-envelope domains (as well as the domain 
at the end of the PTR record) as things that they specifically target 
with their domain/URI lists. But these are things that "consumed" by SA 
with OTHER rules, not with URIDNSBL. (also, postfix as some good rules 
for this too which don't require callouts to content filters like SA. 
Exim and others probably do, too?


At invaluement - we're very very late to this game - and we're going a 
different route - choosing to target these with a separate list, not our 
URI list - this will be our SED list, which is currently under 
development - although, in the meantime, many of our subscribers use our 
existing URI list in this way, outside of our recommendations, and are 
happy with those results.


The main takeaways are:
(1) these require different rules than the URIDNSBL module (since 
URIDNSBL is for checking domains/IPs inside the clickable links in the 
body of the message)
(2) Any DNSBL trying to do should to pay attention to authentication, 
and not just throwing every such domain in the list without being sure 
it really is them and not a forged domain.


I hope this helps!

Rob McEwen, invaluement



Re: Seeking dhl.com ham samples

2022-08-03 Thread Rob McEwen
I provided a ham sample off-list. Also, I've recently encountered a 
similar issues with DHL - for example - them, several weeks ago, using 
an alterate domain in the mail header FROM-address - that didn't 
actually have ANY DNS records - crazy stuff like that - although I think 
that they've since stopped using that particular domain name?

--Rob McEwen

On 8/2/2022 10:50 AM, Bill Cole wrote:
Bug 8021 reports breakage in SPF checking for dhl.com mail, due to an 
inability to resolve the  SPF TXT record for dhl.com. That breakage is 
essentially due to DHL having far too many TXT records (some are 
clearly stale) and having a SPF record which is right at the limit of 
complexity, having 10 'include' directives at the top level.


If anyone has samples of real legitimate mail from a dhl.com address, 
please share. I'm seeking a way to reproduce the reported bug, which 
strikes me as too stupid to be real; we SHOULD have noticed long 
before now if SPF lookups were not handling UDP truncation of replies.




--
Rob McEwen, invaluement



rules for a sneaky SPEAR-VIRUS spam that gets past bayes

2022-03-03 Thread Rob McEwen
rules for a sneaky SPEAR-VIRUS spam that gets past bayes because legit 
content from hijacked emails are copied into the spam, making it look 
like a follow-up msg of an existing legit conversation. Catch using 
these rules below. (Perhaps also add more to this to prevent rare FPs? 
But this is a good start!)


FILE SIZE < 50kb

then, on decoded/demime'd msg:

exact match on:
*https://onedrive.live.com/download?cid=**
*
Then a hit on THIS RegEx:
*\b(Fil lösenord|File password): [A-Z]{2}\d{4}\b**
*

(I'll let someone else jump in here and create and share the actual SA 
implementation of this, if desired - along with any suggested improvements)


-- Rob McEwen, invaluement


Re: Do these domains merit blocking?

2021-12-15 Thread Rob McEwen

On 12/15/2021 11:39 AM, Bill Cole wrote:
Am I missing something special that makes such research spam somehow 
not spam?



Everyone thinks that their own unsolicited bulk email - isn't spam. But 
a line must be drawn somewhere. In this case, the sender has absolutely 
no preexisting relationship to the recipient, and Raymond's statement 
about them sending to "scraped addresses" is, imo, devastating to their 
case. The closest argument that might have been possible is the idea 
that the email might potentially be of more benefit to the recipient 
than it is to the sender (e.g., sort of like a notification about a 
class action lawsuit) - but I can't find that argument anywhere in this 
situation either. But even class action lawsuit notifications are rarely 
sent to scraped addresses.


It's on my "to do" list to add those domains as permanent additions to 
invaluement's URI/domain bl sometime this week, when I get some more 
time. (I'm in the middle of some intense upgrades, so I barely had time 
to type this message.)


-- Rob McEwen, invaluement


Re: OT: is sorbs.net sleeping ?

2021-04-10 Thread Rob McEwen

On 4/10/2021 6:55 AM, Jared Hall wrote:
Rob, I gotta say that I am impressed with the whole Spamhaus-dqs 
program and their use of customer keyed DNS zone queries.  Seems to be 
the way around the client DNS forwarder issues.  How are you guys at 
Invaluement tracking in that area?


I'm not sure I'm understanding what you're saying? Are you referring to 
the fact that their paid customers doing direct queries (NOT the free 
stuff!) - use zone names that have a unique key embedded into the actual 
zone - so that the queries can then be distinguished by this unique key? 
- thus eliminating the need to use the client's local DNS servers' 
public IP as the method of allowing/denying direct queries? Is that what 
you're referring to?



Seems to be the way around the client DNS forwarder issues


If I'm correct about what you meant - then yes - this eliminates 
problems that used to happen when trying to track customers, and 
permission, by IP - because when tracking by an embedded code - then it 
doesn't matter from WHERE the queries come - and queries that come from 
public DNS servers (8.8.8.8 or 1.1.1.1) - can be distinguished one from 
the other - whereas when not doing this - it's impossible to tell 
distinguish the queries from each other and know who is doing them. This 
became especially important because so often the default caching DNS 
server gets auto-flipped to 8.8.8.8, sometimes without the IT person's 
knowledge! And many IT people think that pointing to 8.8.8.8 is the 
textbook way to setup DNS - and have never even heard of things like BIND.


Is THAT what you're talking about?

If so, at invaluement, we've been doing this for 3 years now - but we 
still have a lot of work to do in migrating many long-time customers 
over to our new system. And it was developed before I even knew that 
Spamhaus was doing it this way, and  this involved some extremely 
complex custom modifications of rbldnsd (I couldn't afford to hire an 
expensive high-quality C++ programmer at the time - so it took me about 
100 hours of very intense programming to do that! It didn't help that 
I'm not very good at C++!). I'm not even sure when Spamhaus started this.


Our new system for doing this now involves 86 servers in 43 cities 
around the world - which enables our clients to get their queries 
answered much faster due to accessing an invaluement DNS server with an 
extremely close geolocation. Queries then tend to get answered in a very 
low number of milliseconds - often <10ms.


-- Rob McEwen https://www.invaluement.com +1 (478) 475-9032



Re: OT: is sorbs.net sleeping ?

2021-04-09 Thread Rob McEwen

On 4/9/2021 10:34 AM, Benny Pedersen wrote:
above ip is not listed yet, with inho is sign of no maintain at all 
anymore



So I noticed that this IP you mentioned is a heavily-listed IP that is 
currently listed on many DNSBLs, including many of the best and most 
reliable and accurate ones. (I think that was part of your point.) So 
you're complaining that SORBS isn't listed this one. Maybe you were 
providing this as a representative example, correct? So I guess you're 
saying that there are more like this?


But for the sake of clarity, let me just say that no DNSBLs should ever 
be judged too harshly for "false negatives" - no DNSBL has the exact 
same view of the worldwide email data - and each DNSBL's false positive 
prevention filters will always make SOME mistakes that cause "false 
negatives" - that's a very acceptable price to pay considering that no 
system can ever be perfect.


Low false positives AND overall catch-rates AND overall UNIQUE 
catch-rates (blocking stuff everyone else is still missing) - are all 
far more important metrics.


(you might be disappointed with SORBS in those areas too? - that's fine 
- I'm just trying to clarify that overly judging a DNSBL based on 
/*particular*/ false negatives can be overly harsh and might miss the 
good things that a DNSBL has to offer)


-- Rob McEwen, invaluement +1 (478) 475-9032



Re: Bypass RBL checks for specific address

2020-12-22 Thread Rob McEwen

On 12/22/2020 6:56 PM, Grant Taylor wrote:
Is there a way to bypass RBL checks for a specific address? I've tried 
the all_spam_to option, but it looks like it artificially lowers the 
score and still runs normal tests. I'd like to disable RBL checks for 
one address.



Grant,

First, I'm NOT an expert on all of this - so somebody might be able to 
follow up with BETTER information, but this will hopefully point you in 
the right direction.


So at some point, I think 10+ years ago, I needed to do this, and 
instead of doing this at the spam filter level - I found that it worked 
well to do this via BIND - so this works if you're using your own 
locally-hosted BIND server for resolving DNSBL queries. So, for example, 
if you qualified for use of the free version of SpamCop (as an example), 
and you wanted to whitelist the IP 1.2.3.4, you could add the following 
to your named.conf file:


   zone "4.3.2.1.bl.spamcop.net" in { type master; notify no; file
   "master/null.zone"; };

Then add another such record for each DNSBL that you use. Then the 
"master/null.zone" file can be the following EMPTY zone (or something 
like this - change as desired! This might not even be the best or 
"right" way to do this - this is just what I had in that null.zone file 
when I was doing this 10+ years ago):


   $TTL 52w
   @   IN  SOA root.localhost.  root.localhost (
    2005012001 ; serial
    52w ; refresh
    52w ; retry
    52w ; expire
    52w ; ttl
    )

   @    IN  NS  localhost
   localhost IN    A 127.0.0.1


So by referencing an empty zone file, that way, each "zone" entry just 
points to this one file, for maximum efficiency and caching, and you 
don't have to reenter this for each zone. In the named.conf file, I 
think each zone statement (my 1st example abovw) would go AFTER the 
"options" section, but before the "includes" section. Obviously, if you 
suddenly needed to do this for 10s of thousands of IPs or hostnames, 
then maybe it would then start to have resource/maintenance issues - but 
on a smaller scale, this works great!


HOWEVER - This was so long ago, that I don't know if this STILL works in 
more recent version of BIND without causing issues? It is possible that 
in more recent version of BIND, DNSSEC might interfere with being able 
to do this?


So ANOTHER option might be to use the newer "response-policy" feature - 
my first idea was a sort of hack - but this "response-policy" might be 
more intended for exactly this purpose. So do a search on the following:


"response-policy" "BIND" "NXDOMAIN"
...with each in quotes, as shown, for instructions on how this is done.

So I think the example above, if implemented using "response-policy", 
would be the following:


   response-policy {
  4.3.2.1.bl.spamcop.net IN CNAME .
   };

Or something like that. Double-check my syntax. It might be wrong! So, 
again, I've never done the "response-policy", so this is just to get you 
started and point you in the right direction. If someone comes along and 
corrects my possible mistakes, or provides BETTER info - that is 
excellent - in the meantime, hopefully this will point you in the right 
direction, or give you some ideas.


-- Rob McEwen, invaluement



Re: Mailchimp support for spamassassin-esp

2020-12-01 Thread Rob McEwen

On 11/30/2020 5:40 PM, Alex wrote:

I happened to notice today that the sendgrid spam work being done by
Invaluement (https://www.invaluement.com/serviceproviderdnsbl/) and SA
developers now apparently supports compromised Mailchimp domains.
https://github.com/bigio/spamassassin-esp

Is there an ongoing list of compromised mailchimp domains available to
be used with this? That info is not included with the man page for
this plugin.

I also know there's another plugin developed by Paul Stead for this,
but has one yet become the defacto version yet?



Alex,

So yes - this one *is* the official/defacto version for SpamAssassin. 
This one was developed in coordination with the Apache Foundation and 
its development was partially funded by invaluement. The problem here is 
that the main developer (to his credit - this is a GOOD problem!) got 
ahead of us with the implementations. But we're in the process of 
catching up on the data-generation side, and hope to have those new 
types of data released in the next few weeks (for those ones mentioned 
in those rules, and for other ESPs that will get into those rules 
eventually).


The entire process of developing the engine that produces that SendGrid 
data - was the equivalent of our entire invaluement staff taking at 
least a full month of paid leave away from our regular duties. So that 
got us horribly behind on other things - including getting this data 
into our regular paid datafeeds with instructions sent to our customers 
for that - so we're still catching up on all of that - but we hope to 
get past that soon and to also have those /*other*/ related datafeeds 
for our "service provider DNSBL" released soon. (it won't be as much 
work for the other ones, now that the sendgrid anti-spam data engine is 
already completed - it "blazed the trail") So that explains why a few 
months have passed since the sendgrid data was released without any 
additional data being released yet, and how/why the developer of the 
rules was able to get ahead of us. (again, to his credit!)


Thanks for your patience and understanding.

--
Rob McEwen
https://www.invaluement.com



Re: Apache SpamAssassin and Spammers 1st Amendment Rights

2020-11-20 Thread Rob McEwen

On 11/20/2020 4:59 PM, Charles Sprickman wrote:
non-profit stuff in the US is EXEMPT from CAN-SPAM, so they don’t even 
have to play by the rules



still, that doesn't give them a free pass to the inbox, or prevent them 
from potentially getting listed on an anti-spam list. MANY spammers that 
are blocked by spam filters and/or listed on anti-spam lists - were 
already CAN-SPAM compliant. Being *legal* is a very low bar for email, 
especially in the U.S.


--
Rob McEwen, invaluement



Re: Apache SpamAssassin and Spammers 1st Amendment Rights

2020-11-20 Thread Rob McEwen

On 11/20/2020 4:37 PM, Eric Broch wrote:
It seems spammers are using political arguments to justify their 
actions. I'll give them credit, at least they're trying to justify 
what they do by something greater than (outside of) themselves, albeit 
wrongly.
It seems people on this side of the argument want to jettison politics 
(and religion) and have no justification (only personal preference) 
for what they do. Curious!
At the core spammers seem more logically consistent than those who 
oppose them.



I have extremely large amounts of spams on file in my spamtrap spam 
collection from all various political viewpoints, political parties, and 
moral/ethical/religious viewpoints - MANY of them think that THEIR 
greater good justifies spamming, and ironically their beliefs are often 
in 100% contradiction to OTHER spammers who have opposite beliefs, but 
likewise think that their spam is justified by THEIR "greater good". 
Thankfully, it isn't my job to determine who is justified and, instead, 
I believe that NONE of them are justified in sending spam - spam is 
about *consent* - NOT *content*.


--
Rob McEwen, invaluement



Re: Crap getting through

2020-11-08 Thread Rob McEwen

Daryl,

Can you please post a copy of the raw email message - with headers - 
perhaps with your own user's email address (and name?) masked out 
(change to "") - to pastebin, or to a similar site - then reply 
here with the link. It is difficult to give specific suggestions without 
having the raw underlying text of the message (w/headers). But please 
try to avoid pasting that directly to this list. Thanks!


Rob McEwen


On 11/8/2020 5:00 PM, Daryl Rose wrote:
I'm getting obvious phishing attempts. This one was made to look like 
it was from Wells Fargo with an obvious spoofed email address.  
However, when I examined the headers, the From Address was this 
garbage: *=?utf-8?B?V+G7hWxsc+G4nmFyZ28gQmFuaw==?= *


I received another one that was meant to be an Amazon Prime Membership 
failure.   How can I block these?  The last time I inquired about 
phishing, it was suggested to install KAM, which I did, but this crap 
is still getting through.  Any other suggestions?


Thank you.

Daryl





--
Rob McEwen, invaluement



Re: Invaluement sendgrid list

2020-10-13 Thread Rob McEwen

Micah,

I haven't heard any other complaints about this - so hopefully the 
problem you described is rare or at least intermittent. But, 
nevertheless, I made a change that should GREATLY improve things. Please 
let me know if this fixes/improves/worsens the situation you described. 
Thanks for the feedback - and feel free to continue this conversation 
off-list since the SA list isn't suppose to be the invaluement support 
list. (or, email me at any time about such things - r...@invaluement.com) 
- Thanks!


Rob McEwen, invaluement.com

On 10/13/2020 12:56 PM, micah anderson wrote:

Hi all,

I've been trying the
https://www.invaluement.com/spdata/sendgrid-id-dnsbl.txt list but
lately, I've been getting 'Couldn't connect to server' errors, fairly
regularly. The site says:

'can set them up for frequent downloads (every minute!) using CURL or
WGET - only using the setting that only downloads when the server
versions are newer.'

I am doing that, once per minute... are others having this issue?

thanks



--
Rob McEwen, invaluement
 



Re: blacklisting the likes of sendgrid, mailgun, mailchimp etc.

2020-09-18 Thread Rob McEwen

On 9/18/2020 6:38 AM, Loren Wilton wrote:
https://krebsonsecurity.com/2020/08/sendgrid-under-siege-from-hacked-accounts/ 


also sheds light on the issue too.


. SendGrid knows (or should konw) that it has compromised 
accounts. It could find out what some of them are for free by 
downloading Rob's list of 25 or so compromised accounts.



I strongly suspect that many of those accounts on our 2 Sendgrid lists 
are just plain 'ol spammers, NOT compromised. So some are compromised, 
some are spammers. And the list has grown to 594 SendGrid IDs 
(currently, as I type this) - much more than 25! Also, the list of 
domains found at the end of the SMTP-FROM that we're also deeming as 
spam or malicous has likewise grown to 87 domains


SEE:
https://www.invaluement.com/spdata/sendgrid-id-dnsbl.txt

AND:
https://www.invaluement.com/spdata/sendgrid-envelopefromdomain-dnsbl.txt
https://www.invaluement.com/spdata/sendgrid-envelopefromdomain-dnsbl-rbldnsd.txt
(2nd one formatted for rbldnsd)

I'm seeing evidence/reports that Sendgrid is likely using this data to 
greatly improve their system, and that this (maybe combined with their 
other efforts?) is finally starting to improve things? So that is good 
news. But I'm also shocked at how many hours go by where certain 
egregious accounts on our Sendgrid DNSBLs STILL stay in circulation 
while continuing to send spams, sometimes criminal phishing spams. But I 
also understand that they have to be careful about overly trusting 3rd 
party data, to ensure that they don't overreact to what might be an 
occasional false positive. It shouldn't be too long before they figure 
out that False Positives in those two Sendgrid lists are very very 
rare... practically non-existent. They probably should at least PAUSE 
campaigns pending further investigation. They should at least do that 
much, imo.


(They MIGHT also be suffering from the increasingly common and flawed 
view in the ESP industry - that not-illegal and CAN-SPAM-compliant mail 
is always legit and not spam - mistakenly not understanding that spam 
doesn't have to be illegal and malware, in order to be unsolicited and 
undesired by the recipient (aka "spam"). Maybe them seeing those types 
of accounts in our data is confusing them? I don't know - but much of 
the ESP industry is in great need of a "reset" - and this data is a good 
first step towards that!)


I was planning to spend much time this past week (1) adding this data to 
my own customer's direct query and rsync feeds, and (2) improving the 
instructions, including providing more specific instructions for adding 
this free version to various MTAs - but all that time got put into 
performance and effectiveness enhancements instead. Therefore, the data 
has greatly improved in just the past few days. New data sources were 
added into the mix - and many others of these spams these that were 
previously getting missed, are now getting caught - and the time from 
such a spam being first received - to that data getting into the list - 
has improved from about 1/2 a minute, to just a few seconds!


-- Rob McEwen invaluement.com



Re: ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-25 Thread Rob McEwen

On 8/25/2020 11:04 PM, John Hardin wrote:
I just wrote something similar to generate a rule, in case for some 
reason you don't want to use a plugin. Let me know if there's any 
interest in it. 


yes - please share!

--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-25 Thread Rob McEwen
Thanks, John Capo, for the suggestions! Honestly, I'm at the end of my 
rope - completely burned out from creating this - desperately needing to 
catch up in other areas of my business so that I can pay my bills. And I 
have other ideas for how to make this data even better that I'm trying 
to get to asap. So help like this is very appreciated!


BTW - does Postfix "know" to refresh the data when the files are 
updated? Or is there some kind of command that needs to run to tell 
Postfix to reload the files? How does that work? ALSO - would it help if 
I created a separate set of files for Postfix that are pre-formatted 
this way already?


Thanks!

Rob McEwen, invaluement.com


On 8/25/2020 2:26 PM, John Capo wrote:

On 2020-08-25 11:42, Matus UHLAR - fantomas wrote:


well, do we have anything available now to block at SMTP level?
- postfix policy server?
- milter?

so far I have noticed only SA plugins. Which is not bad, but that HUGE
advantage is not usable now.


Nothing elegant about this but it was easy to implement. You need to 
create the software specific to your MX servers to update the files 
below from Rob's web site.


Adjust the paths below to your Postfix install

Add these entries to your main.cf:

smtpd_restriction_classes =
   sendgrid

# Limit senders that are matched with the regexes in sendgrid-ids
#
sendgrid =
    check_sender_access pcre:/usr/local/etc/postfix/maps/sendgrid-ids

smtpd_recipient_restrictions =
    check_sender_access hash:/usr/local/etc/postfix/maps/from-sendgrid

Create a file like this from the senders in 
https://www.invaluement.com/spdata/sendgrid-envelopefromdomain-dnsbl.txt


sendgrid.net    sendgrid
appliedaicourse.com sendgrid
bithumbcorp.email   sendgrid
bitline.life    sendgrid
bureausveritas.com  sendgrid
caractere.ro    sendgrid
craftsgenerals.com  sendgrid
dalvry.com  sendgrid
...

Name it from-sendgrid and place it in your Postfix directory
postmap from-sendgrid

Create a file like this from the ids in 
https://www.invaluement.com/spdata/sendgrid-id-dnsbl.txt


/^bounces\+2191708-[0-9a-f]{4}-/ REJECT Phish from compromised 
Sendgrid account
/^bounces\+4227563-[0-9a-f]{4}-/ REJECT Phish from compromised 
Sendgrid account
/^bounces\+13780591-[0-9a-f]{4}-/ REJECT Phish from compromised 
Sendgrid account
/^bounces\+10163588-[0-9a-f]{4}-/ REJECT Phish from compromised 
Sendgrid account
/^bounces\+10180020-[0-9a-f]{4}-/ REJECT Phish from compromised 
Sendgrid account

...

Name it sendgrid-ids and place it in your Postfix directory

postfix reload

John Capo
Tuffmail.com



--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-25 Thread Rob McEwen

On 8/25/2020 2:29 PM, Benny Pedersen wrote:
maybe make clamav sigs ? 



Benny,

Thanks for your other suggestions - those are worth exploring.

Also - the Clamav Sigs is not a bad idea - but even besides the fact 
that (like SA rules), Clamav is content filtering and not at the 
SMTP-Envelope level - Clamav doesn't tend to have nearly AS fast of a 
turnaround time as do DNSBLs.


In a previous message, someone was disappointed that we missed one, and 
it turns out our 24-second turnaround time on that message (from the 
start of the SMTP connection - to being fully deployed in the data) was 
a contributing factor. We now have a plan to shorten that 24-seconds to 
about 4 seconds AND (for invaluement subscribers) - we have a "push" 
technology that is available now where those invaluement subscribers who 
opt for this feature (no extra charge!) - can get a split second 
notification to run their RSYNC just 1 second after the file updates - 
and we do that already for our direct query servers. So there is an 
option (once implemented!) to potentially get the these FULLY 
DISTRIBUTED within about 8 seconds from the start of the SMTP connection 
of the first such spam received - to being FULLY deployed on DNS servers 
(both our own direct query servers - and our RSYNC subscribers' internal 
rbldnsd servers) - that will be AMAZING. I expect to be there within a 
week from now. Something like clamav just can't even begin to compete 
with that fast of a turnaround. But ClamAv rules may still be a good way 
to get this implemented for many.


Someone else mentioned one that was completely off of our radar - but 
we're about to double the coverage of these in terms of mailboxes and 
traps used for this purpose - so that will help further minimize our 
"blind spots".


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-25 Thread Rob McEwen

On 8/25/2020 1:20 PM, Rob McEwen wrote:

but I can do everything, at least not all at once


*can't do

--
Rob McEwen
https://www.invaluement.com
 



Re: ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-25 Thread Rob McEwen

On 8/25/2020 11:42 AM, Matus UHLAR - fantomas wrote:

well, do we have anything available now to block at SMTP level?
- postfix policy server?
- milter?
so far I have noticed only SA plugins. Which is not bad, but that HUGE
advantage is not usable now. 



And likewise - 48 hours ago - a SpamAssassin plugin didn't exist either! 
These things take at least a little bit of time. We're only at the 3rd 
business day that this tech has been in existence. But I think you and I 
would both be surprised at how many systems are likely already (quietly) 
using this at the SMTP-connection level, for certain more 
custom-programmed systems. I believe adaptation in other public MTAs is 
inevitable. For example, I have some good contacts at Exim and it's on 
my "to do" list to ask them about this, but I can do everything, at 
least not all at once. And those MTAs that don't enable usage of this 
will be left behind.


PRO TIP: Instead of complaining about this problem on this thread - why 
not go to the discussion list or forum of your preferred MTA - and ask 
them to implement it?


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-22 Thread Rob McEwen

On 8/22/2020 3:35 PM, Kenneth Porter wrote:
--On Saturday, August 22, 2020 11:15 AM -0400 Jered Floyd 
 wrote:



Like most ISPs, they have a feedback loop to remove malicious users.  I
assume it is too slow, so a SendGrid account ID RBL would provide
meaningful value.


Would not Pyzor accomplish the same thing? Submit the SendGrid spam to 
Pyzor to quickly get it blacklisted.



(1) Pyzor requires resource-expensive content filtering - whereas the 
sendgrid list can do the filtering at the SMTP-envelope level - BEFORE 
the message is even downloaded - for some systems with millions of users 
- that is a HUGE advantage.


(2) being filterable at the SMTP-Envelope level opens up possibilities 
for things like MTA plugins or feature additions - that enable this 
filtering at the MTA level - for MTAs that do NOT try to do any content 
filtering of the message. That creates more options for deployment where 
many will hopefully be able to make use of this, who don't have Pyzor 
(for whatever reasons)


(3) The strategy you described is SOMETIMES easily defeated with certain 
variations in the messages, where each message is sufficiently different 
to NOT be blockable by Pyzor. That is a HUGE loophole in Pyzor 
technology. This Sendgrid ID list doesn't have that problem.


(4) Also, a spammer who sends out many different types of spams - can 
potentially stay off of Pyzor's radar - but yet ALL of those spams under 
that Sendgrid ID - will be collectively noticed in our engine. And, 
likewise, Pyzor's methods could create a game of whack-a-mole. The 
spammer will just keep coming out with new types of spam - that all get 
past Pyzor while Pyzor tries to catch up - then Pyzor catches up - then 
the spammer just reformats the content. Rinse. Repeat. Meanwhile, ALL of 
those LATER spams are ALREADY blocked by our Sendgrid list BEFORE the 
next types of spams are sent - ALL OF THEM. (you could argue that we 
might get into a game of whack-a-mole too with those Sendgrid IDs - but 
we're FAR less vulnerable to that - it will happen MUCH LESS often!)


(5) for these reasons and others - I strongly suspect that our Sendgrid 
list is going to have a MUCH faster turnaround time on listing the 
initial spams from a new sendgrid ID - and, as mentioned, their later 
spams will then ALREADY be caught by this Sendgrid list - while Pyzor is 
bogged down in that silly whack-a-mole game.


Don't get me wrong - Pyzor and other such checksum content filters - are 
wonderful and have their place - but thinking that they remove the need 
for this Sendgrid list - is absolutely not even close to true.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one for Sendgrid-spams!

2020-08-21 Thread Rob McEwen
ANNOUNCEMENT: The NEW invaluement "Service Provider DNSBLs" - 1st one 
for Sendgrid-spams!


...a collection of a new TYPE of DNSBL, with the FIRST of these having a 
focus on Sendgrid-sent spams. AND - there is a FREE version of this - 
that can be used NOW! (/well... might need a SpamAssassin rule or two! 
Your help appreciated!)/:


INFO AND INSTRUCTIONS HERE:

https://www.invaluement.com/serviceproviderdnsbl/

This provides a way to surgically block Sendgrid's WORST spammers, yet 
without the massive collateral damage that would happen if blocking 
Sendgrid domains and IP addresses. But we're NOT stopping at the phishes 
and viruses - and we're not finished! There will be some well-deserved 
economic pain, that puts the recipients' best interests at heart. 
Therefore, flagrant "cold email" spamming to recipients who don't even 
know the sender - is also being targeted - first with the absolute worst 
- and then progressing to other offenders as we make adjustments in the 
coming weeks.


-- Rob McEwen https://www.invaluement.com



Re: Bombard by spam source in India that wasn't in any RBL used by spamassassin.

2019-11-06 Thread Rob McEwen

fwiw - this has been blacklisted at invaluement for days.
--Rob McEwen, invaluement.com

On 11/6/2019 2:33 PM, Mark London wrote:
Hi - We got several hours of spam from the IP address 103.136.41.36 in 
India.    When I did a Multi-RBL check, the ip address was in the 
following databases:


bl.emailbasura.org
dnsbl.sorbs.net
dns.spfbl.net
spam.spamrats.com
truncate.gbudb.net

I think sorbs.net is a paid for service.  At least I tried adding 
rules, but they weren't triggered.


I was able to successfully add rules for spamrats and gbudb. Does 
anyone have experience with those?


After about 3 hours, the IP address finally appeared in 
barracudacentra.org, which spamassassin uses.


Given the amount of traffic we were receiving, I'm surprised it didn't 
show up sooner on the other RBLs.


Thanks. - Mark



--
Rob McEwen
https://www.invaluement.com




announcement about invaluement (or more like a tease?)

2019-08-25 Thread Rob McEwen

announcement about invaluement (or more like a tease?)

https://www.linkedin.com/feed/update/urn:li:activity:6571558988201148416/

--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




HostKarma status (was Re: How to block mails from unknown ip addresses?)

2019-08-24 Thread Rob McEwen

On 8/24/2019 6:12 PM, Benny Pedersen wrote:

None of the mails is listed at hostkarma.junkemailfilter.com. I also
use junkemailfilter to score spam.
unmaintained now 



(BCC'ed to new HostKarma mgmt)

Not true. It is under /different/ management now. They are struggling a 
little because of the large learning curve of abruptly stepping into 
someone else's shoes on short notice, with lots of proprietary 
code/processes. Some months ago, I talked to the main tech person over 
there and he is a smart/good guy who is making MUCH progress. So it 
looks like HostKarma is going to survive for the long term.


--
Rob McEwen
https://www.invaluement.com




Re: Freshclam Safebrowsing enabled for SA

2019-04-23 Thread Rob McEwen

On 4/23/2019 11:07 AM, Kevin A. McGrail wrote:

I was going to try and run a second daemon or look at hits for
Safebrowsing. as a method for scoring, not blocking.  The
listing and delisting policies are unclear to me and I think there is a
good potential for FPs.



Probably a nice scoring option - So like Kevin, I'd caution against 
using this for blocking or high scoring. Why? Because in recent years 
there has been an epidemic of the following two things:


(1) website compromised - hacker installed malicious content

(2) email account on the mail server compromised - spammer is sending 
email from that server


HOWEVER - MOST of the time ONLY 1 of these things happened, NOT both. 
But the Safebrowsing database is mainly focused on the website being 
compromised. Therefore, this rule is likely fantastic when it comes to 
hits on content in the body of the message, particularly URLs linking to 
malicious content on hijacked websites. But if/when this instead has 
hits on things like ONLY domain name (in the FROM address or elsewhere) 
- then it might cause a significant number of FPs if/when it hits stuff 
like that.


I'm not very familiar with how this works when implemented in ClamAv - 
so, for example, if this only has hits on entire URLs going all the way 
to the malicious content (not merely referencing the domain or home 
page) - then my FP concerns are likely overstated and this really isn't 
going to cause many FPs.


So I'm just mentioning this so others will be aware and know what to 
look for when testing this.


--
Rob McEwen




Re: How to deel with time limit exceeded

2018-11-05 Thread Rob McEwen
Another thing that helps - is to lighten the load on your SA by putting 
high quality low-FP DNSBLs in front of SA, that are first called by your 
MTA, where spams blocked by those aren't even scanned by SA.

--Rob McEwen

On 11/5/2018 2:48 PM, Andreas Thienemann wrote:


Hi,

I've got a mailserver dealing with a moderate amount of mails... Say 
about 20k mails a day.


The setup has a spamassassin 3.4.2 with a mysql database for bayes and 
to do a minimum of logging.


I've recently seen a lot of mail scans not completing due to running 
out of time. The message I am seeing most often is "check: exceeded 
time limit in 
Mail::SpamAssassin::Plugin::Check::_eval_tests_type11_prineg90_set3, 
skipping further tests". This is happening for abour 5% of incoming mail.


I am trying to understand what exactly is going on here?
It seems that this is happening since I moved bayes to a local mysql 
database to resolve locking problems I saw in the logs with the 
on-disk files. I previously had configured spamd to run with 
--timeout-child=30 which now seems too small.


Having looked at the code of SA I understand that this string is built 
using the following variables:

$package_name."::_".$ruletype."_tests_".$clean_priority;

But that does not really help me understand _what_ is taking that long 
and how I can improve the spamassassin runtime.


Any ideas?

thanks,
 Andreas



--
Rob McEwen



Re: FPs on FORGED_MUA_MOZILLA (for my own hand-typed messages from my latest-version Thunderbird)

2018-10-03 Thread Rob McEwen
The thread has gone somewhat off-topic, which is partly my own fault. 
The issues with URIBL misusage is a "side note", NOT the main purpose of 
this thread. (again, that is party my fault since I mentioned that to 
begin with). Also, I want to make sure that everyone knows that it was 
my client (NOT ME!) that was using URIBL incorrectly. I'll educate my 
client to hopefully fix that problem soon.


NOW... BACK ON THE MAIN TOPIC:

On 10/2/2018 1:52 PM, Matus UHLAR - fantomas wrote:



Message-ID: <39397904-9830-5010-a3d2-a62af8326...@invaluement.com>


this does seem to match:
MESSAGEID =~ 
/^<(?:[a-f\d]{8}-(?:[a-f\d]{4}-){3}[a-f\d]{12}|[A-F\d]{8}\.[A-F1-9][A-F\d]{0,7})\@\S+>$/m


8h-4h-4h-4h-12h@

hmmm we need to look at

(__LYRIS_EZLM_REMAILER || __GATED_THROUGH_RCVD_REMOVER ||
__WACKY_SENDMAIL_VERSION || __IPLANET_MESSAGING_SERVER ||
__HOTMAIL_BAYDAV_MSGID || __SYMPATICO_MSGID)



I really don't think I've done anything unusual with my setup of 
Thunderbird. Does anyone have other suggestions? Is there anything I can 
do with my Thunderbird settings to mitigate this?


Thanks!

--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: FPs on FORGED_MUA_MOZILLA (for my own hand-typed messages from my latest-version Thunderbird)

2018-10-02 Thread Rob McEwen
Bill,

Even though this part wasn't the main purpose of the thread, that is still very 
helpful information. I will pass that along to my client so that they can 
hopefully fix their configuration problem with regards to their usage of URIBL.

Thanks!

Rob McEwen


Sent from my Verizon Motorola Droid
On Oct 2, 2018 11:48 AM, Bill Cole  
wrote:
>
> On 2 Oct 2018, at 9:36, Rob McEwen wrote: 
>
> > SIDE NOTE: I don't think there was any domain my message that was 
> > blacklisted on URIBL - so I can't explain the "URIBL_BLOCKED", but 
> > that only scored 0.001, so that was innocuous. I suspect that that 
> > rule is malfunctioning on their end, and then they changed the score 
> > to .001 - so just please ignore that for the purpose of this 
> > discussion. 
>
> No, "URIBL_BLOCKED" means that the URIBL DNS returned a value that is 
> supposed to be a message to a mail admin that they are using URIBL wrong 
> and will nevewr get a useful answer without either (1) paying for a feed 
> to support their usage volume or (2) using their own recursive resolver 
> instead of forwarding queries to the likes of Google, OpenDNS, & 
> CloudFlare. 
>
> A mail filtering system that gets URIBL_BLOCKED hits is broken. A mail 
> filtering system that gets them chronically is mismanaged. 


Re: FPs on FORGED_MUA_MOZILLA (for my own hand-typed messages from my latest-version Thunderbird)

2018-10-02 Thread Rob McEwen

On 10/2/2018 9:59 AM, Matus UHLAR - fantomas wrote:

can you post the headers?
or at least the Message-Id?



Matus... first, THANKS for your help with this!

Here is the message as THEIR system saw it (with my client's info 
masked)  - but it looks like their Kerio (or the customer's email 
client?) might be not be storing everything as it was originally sent? 
...but this is what my client sent me, fwiw:



Received: from mail.powerviewmail.com 
<http://mail.powerviewmail.com>([204.9.77.40])

by with ESMTPS
(using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256 bits))
for ;
Mon, 1 Oct 2018 15:17:10 +0200
DKIM-Signature: a=rsa-sha256; t=1538399816; x=1539004616; 
s=ivm_invaluement; d=invaluement.com <http://invaluement.com>; 
c=relaxed/relaxed; v=1; bh=C6QzEUsPRf8EoiIEIhSF1hnXxy9JIlmjGFO/079v4QQ=; 
h=From:Subject:Date:Message-ID:To:MIME-Version:Content-Type:In-Reply-To:References;

b=V5Sv2lZUWL4P29pcEVY6r/8uFRcuNL1hR794r6M1TJZcvw+i4vTgrvWf+CKSN/F1f2FS/0CdF4UCux+dS/vFjj3X9fdmwv9jpizZqwvJseyCYEmT2HItdeqo0NfNIoQwziEPDMgYS3f35iWlcb7wqrPjfx5EslHr+oC0eoeGBaA=
Received: from [204.9.77.40] ([204.9.77.40])
        by mail.powerviewmail.com 
<http://mail.powerviewmail.com>(IceWarp 12.0.2.1 x64) with ASMTP id 
201810010916565985

        for ; Mon, 01 Oct 2018 09:16:


Here is an excerpt from the headers, copied from the message in my 
Thunderbird "sent" folder:



References: <55521fa7.8080...@invaluement.com> 
<7c8ad385-8b3d-74d9-7d34-ca2ca9236...@invaluement.com> 
 
<1b8ad5ec-18b7-90db-5cad-d86ffa5aa...@invaluement.com> Message-ID: 
<39397904-9830-5010-a3d2-a62af8326...@invaluement.com> 
Disposition-Notification-To: Rob McEwen  Date: Mon, 
1 Oct 2018 09:16:55 -0400 User-Agent: Mozilla/5.0 (Windows NT 10.0; 
WOW64; rv:52.0) Gecko/20100101 Thunderbird/52.9.1 MIME-Version: 1.0 
In-Reply-To: <1b8ad5ec-18b7-90db-5cad-d86ffa5aa...@invaluement.com> 
Content-Type: multipart/mixed; 
boundary="54AEB3A413950E8E0A41E1A8" Content-Language: en-US




The time difference makes sense because their time zone is 6 hours ahead of 
mine.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




FPs on FORGED_MUA_MOZILLA (for my own hand-typed messages from my latest-version Thunderbird)

2018-10-02 Thread Rob McEwen
A client of mine wasn't getting my own hand-typed messages. 
Unfortunately, they had their SA set to block on a score of 3 (which is 
aggressive), and this particular rule hit plus a tiny bit of other 
things put it above 3. But what is weird - is that it was hitting on 
hand typed-messages from me - that I sent directly from my 
latest-version of Thunderbird. So this was NOT "forged" at all! (Also, I 
suspect that the bayes hit was due to previous such messages from me 
getting blocked and feeding his bayes?)


Any suggestions? Could my client be using a very old version of SA - 
where this is fixed already? (they are using SA from Kerio).


Here are the headers:

X-Kerio-Anti-Spam:  Build: [Engines: 2.15.8.1169, Stamp: 3], Multi: 
[Enabled, t: (0.12,0.017258)], BW: [Enabled, t: (0.13)], RTDA: 
[Enabled, t: (0.052863), Hit: No, Details: v2.7.15; Id: 
15.1i65djr.1conscun2.ocr1k], total: 0(700)

X-Spam-Status: Yes, hits=3.8 required=3.0
tests=KERIO_ANTI_SPAM: -0.000, AWL: -0.000, BAYES_50: 1.567,
FORGED_MUA_MOZILLA: 2.309, HTML_MESSAGE: 0.001, URIBL_BLOCKED: 0.001,
TOTAL_SCORE: 3.878,autolearn=no

Suggestions?

SIDE NOTE: I don't think there was any domain my message that was 
blacklisted on URIBL - so I can't explain the "URIBL_BLOCKED", but that 
only scored 0.001, so that was innocuous. I suspect that that rule is 
malfunctioning on their end, and then they changed the score to .001 - 
so just please ignore that for the purpose of this discussion.


--
Rob McEwen
https://www.invaluement.com




Re: using URIBL on other headers

2018-09-23 Thread Rob McEwen
to limit false 
positives will cause them to defer on the listing - even though it would 
have been an excellent and justified ratio of spam-to-ham blocked, with 
little collateral damage if the mail systems using that list could have 
ONLY blocked using one method or the other, NOT both! DBL likely errs on 
the side of less collateral damage - so it is should be safe to use DBL 
for blocking based on both methods, as they prescribe, especially 
considering Spamhaus' reputation for extremely low false positives. 
Then, other URI lists can pick up the slack on the occasional False 
Negatives.


(7) Given this information, at invaluement, we have solved this problem 
by creating a new domain blacklist ("ivmSED") that is independent of 
ivmURI, where ivmSED is a domain blacklist used ONLY for blocking based 
on the domains found on the SMTP envelope (FROM, PTR record, HELO) - and 
where ivmSED NOT be used for blocking domains in clickable links in the 
body of the message, since that is the job of our ivmURI list. That way, 
ivmSED and ivmURI are independent, and we then have the flexibility to 
block a domain using either method independently, or both together, for 
the approach that most surgically targets the spam, keeps collateral 
damage to a minimum, and without compromises that lead to more false 
negatives. ivmSED has just recently entering beta testing. (SED = 
"Sender's Envelope Domain").


--
Rob McEwen
https://www.invaluement.com




Re: DNS and RBL problems

2018-09-14 Thread Rob McEwen

On 9/14/2018 1:36 PM, Alex wrote:

Hi,

For the past few weeks I've been having problems with queries to many
of the common RBLs, including barracuda, mailspike and unsubscore. My
logs are filled with "Name service error", SERVFAIL and lame-server
messages for RBLs I know to be valid.




Alex,

Coincidentally, a recent new invaluement subscriber was initially having 
at least similar problems that didn't make sense. I was stumped. It made 
no sense that it wasn't working because everything looked correct. But 
then he figured out that the following bug was the cause, and fixing 
this bug enabled the queries to start working again:


NOTICE: SpamAssassin installations affected by a bug, due to a change 
Net::DNS made in an earlier version, here is the bug for reference:

https://bz.apache.org/SpamAssassin/show_bug.cgi?id=7223

So you should definitely check to see if this is causing your problem?

--
Rob McEwen
https://www.invaluement.com




Re: CVE-2018-12558: DOS in perl module Email::Address

2018-06-20 Thread Rob McEwen

On 6/20/2018 1:30 PM, Bill Cole wrote:

http://www.openwall.com/lists/oss-security/2018/06/19/3

SpamAssassin does not use Email::Address.



Thanks, Bill, for clarifying that. I've been concerned about this for 
hours - but too busy today research it myself.


--
Rob McEwen
 



Re: OT: Congratulations Dianne

2018-04-03 Thread Rob McEwen

On 4/3/2018 1:18 PM, Axb wrote:

AppRiver Acquires Roaring Penguin
https://globenewswire.com/news-release/2018/03/26/1453063/0/en/AppRiver-Acquires-Roaring-Penguin.html


Excellent! Dianne, I hope you benefited greatly in this acquisition!

--
Rob McEwen
https://www.invaluement.com




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-04-03 Thread Rob McEwen

On 4/3/2018 9:27 AM, Leandro wrote:
We just created an URL signature algorithm to be able to query an 
entire URL at our URIBL:


https://spfbl.net/en/uribl/

Now we are able to blacklist any malicious shortener URL



Leandro,

Thanks for all you do! And good luck with that. But there are a few 
potential problems. When I analyzed Google's shortners about a month 
ago, I found that a VERY large percentage of the most malicious 
shortened URLs were a situation where the spammers were generating a 
unique shortner for each individual message/recipient-address. This 
causes the following HUGE problems (at least for THESE particular 
shortners) when publishing a full-URL dnsbl:


(1) much of what you populate your rbldnsd file with is going to be 
totally ineffective for anyone since it ONLY applied to whatever single 
email address where the spam was original sent (where you had trapped 
it) - everyone else is going to get DIFFERENT shortners for the spam 
from these same campaigns that are sent to their users.


(2) get ready for EXTREME rbldnsd bloat. You're gonna need a LOT of RAM 
eventually? And if you ever distribute via rsync, those are going to be 
HUGE rsync files (and then THEY will need a lot of RAM). Sadly, most of 
that bloat is going to come from entries that are doing absolutely 
nothing for anyone.


(3) You might be revealing your spam traps to the spammers. In cases 
where the spammers are sending that 1-to-1 spam to single recipient 
shortners, then all they gave to do is enumerate through their list of 
shortners, checking them against your list - and they INSTANTLY get a 
list of every recipient address that triggers a listing on your DNSBL. 
If you want to destroy the effectiveness of your own DNSBL's spam traps 
- be my guest. But if you're getting 3rd party spam feeds (paid or free) 
- then know that you're then screwing over your 3rd party spam feed's 
spam traps - and those OTHER anti-spam system that rely on such feeds, 
which will then diminish in quality. (unless you are filtering OUT these 
MANY 1-to-1 shortner spams)


Maybe there is enough OTHER shortners (that are sending the same 
shortners to multiple recipients) to make this worthwhile? But the bloat 
from the ones that are uniquely generated could be a challenge, and 
could potentially cause a MASSIVE amount of useless queries. I'd be very 
interested to see what PERCENTAGE of such queries generated a hit!


Meanwhile, in my analysis I did about a month ago, about 80% of Google's 
shortners found in egregious spams (that did this one-to-one 
shorter-to-recipient tactic)... were all banging on one of ONLY a dozen 
different spammers' domains. Therefore, doing a lookup on these and then 
checking the domain found at the base of the link it redirects to... is 
a more effective strategy for these - whereas, for THESE 80% of 
egregious google shortners, a full URL lookup is worthless, consuming 
resources without a single hit.


Alternatively, you may have found a way to filter out these types of 
individualized shortners, to prevent that bloat? But even then, everyone 
should know that while your new list might be helpful, it would be good 
for others to know your new list isn't applicable to a large percentage 
of spammy shortners, since it is still useless against these 
individualized shortners.


NOTE: Google has made some improvements recently, and I haven't yet 
analyzed how much those improvements have changed any of these things 
I've mentioned?


PS - the alphanumeric code at the end of these shortners tend to be 
case-sensitive, while the rest of the URL is NOT case sensitive (and 
they also work with both "https" and "http")... so you might want to 
standardize this on (1) https and (2) everything lower case up until the 
code at the end of the shortner - before the MD5 is calculated. 
Otherwise, it could easily break if the spammer just mixes up the 
capitalization of the shortner URL up until the code at the end of the 
shortner.


--
Rob McEwen
https://www.invaluement.com



Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-04-01 Thread Rob McEwen

On 4/1/2018 7:10 PM, Kevin A. McGrail wrote:
No, I don't think it's an April Fool's trick though it is possible. 
They announced this a day or 2 ago.
See 
https://www.cloudconnectcommunity.com/ccc/ls/community/g-suite-feature-ideas/post/6320666165116928 
and https://firebase.google.com/docs/dynamic-links/



I have been in talks with Google about this, sharing with them specific 
data and stats. My contact there sent me an email about this a day or 
two ago referencing that announcement - I just hadn't had time to review 
it yet. So this is for real.


The data I had compiled and shared with them that I gathered about a 
month ago... was devastating. I can't make that report public because 
the spammers were OFTEN sending a one-to-one ratio of a single 
individual google shortener per recipient address. Therefore, making 
that data public would have revealed many spamtrap addresses (both mine 
and those of 3rd party spam feeds). Another problem with that data... is 
that I was finding 10s of thousands of google shortners that were 
banging on the same dozens spammer's domains - and it would be so 
trivial for Google to look for that pattern and then just wipe out all 
of THOSE many shortners that were redirecting to one of these dozen 
spammers' domains. At the same time, the vast majority of these 
shortners were persisting for weeks and weeks with no sign that Google 
was hardly lifting a finger to terminate them. In fact, in that initial 
report, 95+% of shortners used by spammers were persisting for weeks on 
end, and 80% of those were using one of those 12 spammer' domains.


HOWEVER... I started compiling new data few days ago, and it looks MUCH 
better now. I'm not done putting that report together, but early 
indications show that something has recently changed for the better - 
but that there is still significant room for improvement, but they are 
at least headed in the right direction. But that report is still under 
construction.


So I think the pressure on Google, along with some of the data I 
provided them... might have helped? Or maybe that was just "one straw 
that broke the camel's back"? Either way, I'm happy that this seems to 
be getting fixed, or they are at least headed in the right direction.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: sneaky spams w/zipped URL file, easily caught by "Thread-Index"

2018-03-27 Thread Rob McEwen

On 3/27/2018 9:48 AM, David Jones wrote:

Looks like ClamAV UNOFFICIAL sigs are detecting this:
Clamd: message was infected: Sanesecurity.Foxhole.Zip_url.UNOFFICIAL 



David,

Excellent... except for one potential problem... this is in their 
"foxhole_all.cdb" file which they label as "high false positive risk" - 
which could scare some away!


For those who don't score very high on ClamAv and/or who are able to 
score DIFFERENTLY based on different types of Sanesecurity and/or ClamAv 
results, this is probably OK. But for others who prefer to either 
outright block or score high on ClamAv, that MIGHT present a problem. On 
the other hand, maybe Sanesecurity is just being overly cautious (or 
considering more theoretical FNs?), and such actual FPs in real world 
mail flow are actually extremely rare?


Any Thoughts? Anyone know?

--
Rob McEwen
https://www.invaluement.com



sneaky spams w/zipped URL file, easily caught by "Thread-Index"

2018-03-27 Thread Rob McEwen
Today, MUCH sneaky spams are being sent with an attached zipped 
malicious URL/shortcut file.


Most or all of these are easily caught by Thread-Index, as follows:

Thread-Index: AdBx5/5UsdSTxflQTPi+FyODmVaqhA==

Perhaps someone can make a rule for this and post it here?

I already set this in another non-SA part of my anti-spam system, but 
the rule might help others here. There are also other attributes that 
could become an SA rule that would cause a hit even if the Thread-Index 
changed, but that will require a little bit more effort.


--
Rob McEwen
https://www.invaluement.com




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-03-15 Thread Rob McEwen

On 3/15/2018 11:13 AM, sha...@shanew.net wrote:

You might take a look at
https://developers.google.com/url-shortener/v1/getting_started

1 miion requests per day is the default limit.



Excellent! Thanks for the suggestion. This should help me MUCH!

But, unfortunately, it still leaves a lot to be desired for fixing the 
problems I described... when it comes to more distribution of automated 
"plugins" or having spam filters automatically doing this - in a "set it 
and forget it mode".


Still, there is a good argument that if someone has a high enough volume 
to trigger rate limiting... then they ought to have a large enough staff 
to handle going the extra mile and setting up API access systems, etc. 
But I wish it could be more simple. MANY are going to struggle with 
this... and NOT easily know all the details discussed on this thread!


But again, this should help me (and others) much... and it is good to 
know that there is a proper way to do this at a higher volume that meets 
Google's approval.


--
Rob McEwen
https://www.invaluement.com



Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-03-14 Thread Rob McEwen

On 2/20/2018 9:42 PM, Rob McEwen wrote:
Google might easily start putting captchas in the way or otherwise 
consider such lookups to be abusive and/or mistake them for malicious 
bots...


This prediction turned out to be 100% true. Even though others have 
mentioned that they have been able to do high-volume lookups with no 
problems... And granted I wasn't implementing a multi-server or multi-ip 
lookup strategy... But I don't think I was doing nearly as many lookups 
as others have claimed that they were able to do. I took a batch of 
55,000 spams that I had collected from the past 4 weeks where those 
spams were maliciously using the Google shortener as a way to get their 
spam delivered via hiding their spammy domain names from spam filters. I 
started checking those by looking up the redirect from Google's 
redirector, but without actually visiting the site that the redirector 
was pointing to. Please note that I was doing the lookups one-at-a-time, 
not starting the next lookup until the last lookup had completed. After 
about ONLY 1,400 lookups, ALL of my following lookups started hitting 
captchas. See attached screenshot. Also, other than not sending from 
multiple IPs, I was otherwise doing everything correct to make my script 
look/act like a regular browser.


I'll try spreading it out between multiple IPs in order to try to avoid 
rate limits... However... This is still cause for concern about 
high-volume lookups in high production systems... those may have to be 
implemented a little more carefully if they're going to do these kind of 
lookups!


Just because small or medium production systems are able to do this... 
Or just because somebody went out of their way to get more sophisticated 
with it to get it to work out... doesn't mean that it's going to work in 
high production systems that are trying to use "canned" software or 
plugins. This is a particular challenge for anti-spam blacklists because 
they typically process a very high volume of spams. Hopefully, the 
randomness of the ones I process as they come in... will be sufficiently 
spread out enough to avoid rate limiting?


It was my hope to start processing these live with my own DNSBL engine, 
so that I could start blacklisting the domains that they redirect to... 
In those cases where they were not already blacklisted... Now I'm going 
to have to deal with constantly trying to make sure that I'm not hitting 
this captcha, along with implementing some other strategies to hopefully 
prevent that.


But this brings up a whole other issue... That is more of a policy or 
legal issue... is Google basically making a statement that automated 
lookups are not welcome? Or are considered abusive?


(btw, I could have collected order of magnitudes more than 55,000 of 
THESE types of spams, but this was merely what was left over in an 
after-the-fact search of my archives, after a lot of otherwise redundant 
spams had already been purged from my system.)


PS - Once I gather this information, I will submit more details about 
the results of this testing. But what is shocking right now is that less 
than four tenths of 1% of these redirect URLs has been terminated, even 
though they average two weeks old, with some almost a month old.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-03-10 Thread Rob McEwen

On 3/10/2018 11:43 AM, Matus UHLAR - fantomas wrote:

On 3/10/2018 11:22 AM, Matus UHLAR - fantomas wrote:
this is apparently not the case of one url redirector (shortener) 
points to

another shortener.

I really hope that the DecodeShortURLs only checks fopr redirection 
at those

known redirectors (shorteners), not each http->https shortener and only
evaluates redirection between them, ignoring http->https redirects


On 10.03.18 11:32, Rob McEwen wrote:
But also keep in mind that it is NOT rare for the initial shortner 
found in a spam... to redirect to a spammer's page (that isn't a URL 
shortner), then THAT page then redirects to the spammer's OTHER page 
(that isn't a URL shortner)... where the FIRST one is NOT blacklisted 
very many places... but the second page (that is often the final 
destination)... *is* heavily blacklisted. Often, the first page is 
hosted at a legit site that was hacked into by criminal spammers - 
and is HARDER for blacklists to list due to their good reputation. 


isn't this thread and the plugin discussed about shorteners?
I am aware of other possibilities to do redirects and about 
consequencies of

following them.
note that for following the redirect, the destination must be 
configured via

url_shortener directive.
what I wonder is, if there's real value in following the redirects.
(and if the following stops when the URL is blacklisted, since we're done
here).



The URL shortner follower plugin that Steve Freegard wrote... ALREADY 
follows my suggestions about following non-shortner redirects (after the 
initial shortner redirect). (except I'm not sure it already does my 
https suggestion?) ...and for good reason. You're tremendously 
undervaluing or just not understanding my last post. I suggestion you 
re-read it a couple of times. This tactic I mentioned that spammers 
use... is COMMON, and not following my last suggestion opens up a big 
loophole for spammers. Such spammers would be delighted by your last 
comment... and Steve Freegard went a different route for good reason. 
Keep in mind, I haven't veered off-topic because we're STILL talking 
about a possible chain of redirects that was still STARTED by a URL 
shortner.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-03-10 Thread Rob McEwen

On 3/10/2018 11:22 AM, Matus UHLAR - fantomas wrote:
this is apparently not the case of one url redirector (shortener) 
points to

another shortener.

I really hope that the DecodeShortURLs only checks fopr redirection at 
those

known redirectors (shorteners), not each http->https shortener and only
evaluates redirection between them, ignoring http->https redirects 



But also keep in mind that it is NOT rare for the initial shortner found 
in a spam... to redirect to a spammer's page (that isn't a URL 
shortner), then THAT page then redirects to the spammer's OTHER page 
(that isn't a URL shortner)... where the FIRST one is NOT blacklisted 
very many places... but the second page (that is often the final 
destination)... *is* heavily blacklisted. Often, the first page is 
hosted at a legit site that was hacked into by criminal spammers - and 
is HARDER for blacklists to list due to their good reputation. Then the 
2nd final destination page is just a heavily blacklisted spammer's 
throwaway domain. Therefore, there is some value to following the 
redirects a little further than what you suggest, and then collecting 
all of those host names or domains, checking ALL of them against 
URI/domain blacklists. (within reason... after too many redirects, it is 
better to just stop and add points to the spam score)


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-03-10 Thread Rob McEwen

On 3/10/2018 3:20 AM, Matus UHLAR - fantomas wrote:
do you have an example of any chained redirection not suspicious? 



I haven't examined the code for that plugin very much (yet!) but one 
type of very common redirect that is very innocent... is the fact that a 
MASSIVE percentage of web sites have switched to all SSL sites in the 
past few years, due to Google now punishing http and/or rewarding 
https... in the search engine rankings. But there are still many links 
and redirectors pointing to the non-SSL versions of those sites, which 
is then typically redirected to the SSL version. Therefore, if the code 
for this plugin (and others using this tactic) doesn't do this 
already... it should probably not count THAT particular redirect as a 
spam indicator, when counting the total number of redirects.


--
Rob McEwen
https://www.invaluement.com




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-27 Thread Rob McEwen

On 2/26/2018 1:00 PM, Kevin A. McGrail wrote:

DecodeShortURLs has been on my list of must-have plugins for years, so
I was a little surprised it took so long for someone to mention it
in this thread. 


Yeah, my firm is going to look at subsidizing it's addition to SA and 
Karsten agreed.  Just trying to get the time.




Kevin,

Excellent! Also... for those situations where there are multiple 
cascading redirecting websites (after the initial URL shortners link 
found in the message), please consider having an option for *every* 
non-whitelisted URI in the chain to be added to the checks against 
blacklists. OFTEN - every single domain in that chain (past the initial 
URL shortner) is a compromised web site or spammer's website, not just 
the final destination web site.


--
Rob McEwen
https://www.invaluement.com
 



Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-21 Thread Rob McEwen

On 2/21/2018 2:41 PM, Amir Caspi wrote:

On Feb 21, 2018, at 9:57 AM, Dianne Skoll <d...@roaringpenguin.com> wrote:

That's why you only want to do it for URLs that are
absolutely known to be shortened URLs.  You have to keep a list of
known URL-shorteners.

On that note -- regardless of what OTHER HW/SW solutions might do, since this 
is a SpamAssassin mailing list ... is there any facility to implement this in 
SA?  That is, when calling the URIBL plugin, could it check both the shortened 
URL and the expanded URL (for known shorteners) ?  Does that facility already 
exist and I missed it?


If I'm not confusing things, someone answered things earlier in this 
thread, as follows:


On 2/21/2018 11:27 AM, Alex wrote:

This is what DecodeShortURLs is for
https://github.com/smfreegard/DecodeShortURLs



--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-21 Thread Rob McEwen

On 2/21/2018 11:48 AM, Anthony Cartmell wrote:

Meanwhile - adding URI lookups (for URIs in the body of the domains)
and/or the option to add 3rd party URI list lookups - is STILL is
missing from MANY widely used anti-spam systems.

If you mean following URLs in messages, you do need to be aware that
this can break one-time login links.

I have come across one local authority in the UK whose spam filter
requests URLs in incoming messages. This meant that the spam filter had
already used the one-time login link so that when the intended user
tried they got an error saying that the link had already been used. Took
a while to work out what was going wrong!



Sorry I wasn't more clear. I was just referring to standard things like 
usage of SURBL, URIBL, ivmURI, and Spamhaus's DBL list - where the 
domains and IPs that are in the body of the message are collected and 
checked against those lists - I didn't mean to imply that such links are 
actually followed. Again, sorry I wasn't more clear. I was just pointing 
out that if so many systems STILL don't do THAT - or provide an 
opportunity for adding additional such lists - then I doubt that we're 
going to see widespread adoption by mail systems of a process where, in 
real time spam filtering, they check to see where URL shortners lead to, 
and then factor that destination into the spam filtering.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: Expanding shortened URLs (was Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response))

2018-02-21 Thread Rob McEwen

On 2/21/2018 11:44 AM, Dianne Skoll wrote:

On Wed, 21 Feb 2018 16:35:27 +
Karol Augustin <ka...@augustin.pl> wrote:

I think the point here might be that if Google acted promptly on abuse
spammers would stop using shorteners.

True, that might happen.  OTOH, I see about as many spams with bit.ly
shorteners as goo.gl shorteners which is not what one might expect if
bit.ly were really that much more proactive than goo.gl.


I'm sure mileage may vary - but I'm seeing about 10X the abuse for 
goo.gl right now as I see from bit.ly. Also, when I do random checks on 
a handful of abused bitly shortners, a high percentage of them are 
already terminated. But when I do random checks of abused goo.gl 
redirectors, most of them are still operational. (I'm referring to 
redirectors found in spams within the previous few days of when I 
checked them, with at least hours having gone by since the message was 
sent - I know that sounds anecdotal - but as I've been researching this 
in the past weeks, that pattern keeps happening). One thing that could 
potentially make those numbers different - is when you compare one 
system that blocks MUCH spam at the perimeter based on the sending IP, 
such as blocking all Zen-listed spams before DATA while another 
system might capture ALL messages and process them all. The latter is 
what my system does. That also might explain the difference in stats?


--
Rob McEwen
https://www.invaluement.com



Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-21 Thread Rob McEwen

On 2/21/2018 11:12 AM, Dianne Skoll wrote:

Really?  This isn't rocket science.  If I thought of it, I'm sure dozens
if not hundreds of others have thought of it and implemented it.


Meanwhile - adding URI lookups (for URIs in the body of the domains) 
and/or the option to add 3rd party URI list lookups - is STILL is 
missing from MANY widely used anti-spam systems. That isn't rocket 
science either - and is MUCH more needed and beneficial than this 
feature. That some others have made adjustments to their system to 
follow and act on redirectors during live spam filtering - again - 
doesn't alter my original point. The vast majority of anti-spam systems 
in the real world (1) don't (2) and won't any time soon. That is what I 
claimed. Please stop nitpicking and please stop arguing with a "straw man".


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: Expanding shortened URLs (was Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response))

2018-02-21 Thread Rob McEwen

On 2/21/2018 11:11 AM, Dianne Skoll wrote:

I guess I misinterpreted: "...such automated lookups could also put a
huge extra burden on Google's servers..." from Message-Id
<b229b274-ff75-0109-9510-346b078d9...@invaluement.com>


Oh yeah, I'd forgotten about that part. it was a more minor point. But 
as I think back on my thought processes at the time I typed those words 
- I was envisioning what would happen if ALL ISPs and hosters and spam 
filtering vendors, SA installations etc...  ALL started doing all of 
those lookups. But yeah, maybe that would still be a drop in the bucket 
compared to all that Google does.


Nevertheless, it is a shame to have to shift more of the burden onto 
spam filters to do more work (some of which requires MORE latency) - in 
order to partly mitigate Google's failure to prevent/correct the abuse. 
In contrast, at this time, bit.ly is running circles around Google in 
terms of quickly shutting down their abused redirectors. I know this 
isn't easy, but there is definitely room for improvement.


But my larger point in that overall post you quoted from, was my concern 
about one organization doing high volume lookups from a single server 
getting blocked or captcha'd.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-21 Thread Rob McEwen

On 2/21/2018 10:39 AM, Dianne Skoll wrote:

We use HEAD requests to expand known URL-shorteners on a cluster that
peaks around 60 msgs/s


Thanks for that information. That is good to know!


(b) and this isn't going to suddenly become a feature inside of many
types of spam filtering hardware and software overnight... that could
even take years,

It's been part of our practice for about a year now.


Excellent! I wish others would be as innovative and on top of things as 
you are! Unfortunately, your statement doesn't alter my point you were 
replying to, even one tiny bit.


--
Rob McEwen
https://www.invaluement.com




Re: Expanding shortened URLs (was Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response))

2018-02-21 Thread Rob McEwen

On 2/21/2018 10:37 AM, Dianne Skoll wrote:

The concern voiced in another email about overloading Google's
infrastructure is quite charming and quaint.



My concern was NEVER about overloading google. My concern was about 
Google auto-blocking or throwing a captcha at very high volume and 
automated lookups. That is a HUGE difference.


--
Rob McEwen
https://www.invaluement.com




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen

On 2/21/2018 1:38 AM, @lbutlr wrote:

As I suspected, it is possible to get the goo.gl target URL without loading the 
site, though using curl is probably not realistic in this specific case.



That is an idea worth exploring! Some might greatly benefit from that.

However:

(a) it might not "scale" for high volume mail flows and DNSBLs who, like 
invaluement, process dozens (or more) spams per second.


(b) and this isn't going to suddenly become a feature inside of many 
types of spam filtering hardware and software overnight... that could 
even take years, if it could ever even gain traction? It is taking into 
the decades just to get some of those software and hardware vendors to 
add support for URI blacklists, or support to for adding custom 3rd 
party URI blacklists. If that is taking literally decades - they are not 
going to add this feature within the foreseeable future.


So please don't think for a second that this somehow makes the plans I 
had described as unnecessary.


--
Rob McEwen
https://www.invaluement.com





Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen
ic if only idiots ran spam filters and DNSBLs. But, 
thankfully, I'm able to separate the legit uses - from those spammers 
who are trying to hide their identities in order to facilitate spamming 
- a point lost on you. I've NEVER ONCE argued for creating rules that 
blocked ALL messages that used shortners - but that is the "straw man" 
that you're arguing against. This is also one of the very reasons that 
spammers are so happy with this loophole - they know that nobody is 
going to create a rule that blocks goo.gl (or even scores that high 
against it). But thankfully, they'll get SOME relief as more of those 
who abuse it find their IPs blacklisted at invaluement (and hopefully 
other places), and rightly so! (yet without the collateral damage you 
keep predicting or implying)


If we're going to continue this conversation, can you please STOP 
putting words in my mouth and arguing against "straw men"? Also, I 
understand your very valid concerns about collateral damage. I've 
addressed that numerous times and in numerous ways, in numerous posts. 
This is getting tiresome.


--
Rob McEwen
https://www.invaluement.com




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen

On 2/20/2018 6:05 PM, @lbutlr wrote:

On 2018-02-20 (08:30 MST), Rob McEwen <r...@invaluement.com> wrote:

Spammers are starting to use this to evade spam filters,

This is not news. Spammers have been using shortness since 3 seconds after 
tinyurl.com launched.


My "this" was /*specifically*/ referring to Google's shortner, or at 
least the recent STRONG uptick in the abuse of that shortner. I was 
already well aware of other similar things from other shortners. But 
"this" wasn't referring to them. You infused thoughts/meaning into my 
writing that really wasn't there. (be careful about assuming...) Also, 
when you see my report further down, you'll understand why those others 
are not nearly as much of a concern to me at the moment.



Keep in mind that, if a marketer is doing things the right way, they should 
have no need to obfuscate their own domain name. They should instead proudly 
use it and not feel the need to hide behind Google's shortner.

No, that is not at all true. The primary use of a shorter is to shorten a long 
URL to something that someone can type in.


I've acknowledged that there are some good reasons for a shortner - but 
the vast majority of the times I'm seeing them - in both ham and spam - 
that is NOT the case! The are shortening things like average-sized 
domains with either no directory, or with a short directory or page 
names after the domain. This is the VAST MAJORITY of the shortners I'm 
seeing in both hams and spams.



Clicking a URL in an email is the height of stupidity, so having a short URL 
that someone can realistically type into a browser is much better.


If I spent just a little more time on this, I could collect a large 
number of Google shortner URLs that are malicious - where my 
malwarebytes is blocking access to the page to which it is trying to 
redirect. And these are still "live"! Do you really think that more than 
a tiny percent of those who saw those in their mailbox (both legit and 
spams) are manually typing in the URL and not just clicking on it? And 
in that exceedingly rare occasion where somebody types in the URL and it 
redirects to a malicious page that tries to install a virus, are they 
ANY better off than having just clicked on it? Even the best point I can 
think of that you might have had - that this might help them to better 
recognize a phishing URL for example - is lost since BOTH the phish and 
their bank's web site is going to be indistinguishable until AFTER 
they've launched the shortner (whether by clicking or typing). I think 
you just mistakenly bolstered my argument against this over-usage (and 
often inappropriate usage) of shortners!



THEREFORE: If you like having NOT-blacklisted IPs, be advised that the invaluement anti-spam DNSBL system is 
now adding "bad" points to the scoring of all messages that use the "goo.gl" shortner, 
and we're amplifying other "bad" points.
Well, at least you are warning people. However, what you are doing is, 
frankly, dumb; if you think there's a huge problem, you can simply 
check the target URLs. 


That incurs a significant amount of extra resources for DNSBLs and spam 
filters - and such automated lookups could also put a huge extra burden 
on Google's servers - and who knows at this point if this is even 
reliable - Google might easily start putting captchas in the way or 
otherwise consider such lookups to be abusive and/or mistake them for 
malicious bots... I'm definitely going to pursue this further - but 
wow... that you would suggest this... I think spam puts ENOUGH burden on 
spam filters and mail system as it is!



Yes, there are many legitimate uses of Google's shortner, too. However, we are 
now at a point where a VERY large % (a majority?) of uses of these headed to a 
typical user's mailbox are egregious spams, and a significant additional 
portion are likely-spams.

Any evidence of this?



EVIDENCE/STATS:

I ran stats on a sample set of a few thousand mailboxes, over a period 
of several hours today (mostly during business hours for these 
particular organizations who use these mailboxes) - and this produced a 
combined 24K legit messages, and 5K spams (I'm guessing that most 
systems have more spams per amount of hams? But those were the numbers 
for this server.)


---
NOTE: The sum of individual shortner-hits totals below can be LARGER 
than the total messages that had hits on ANY shortner - Why? - Because 
in a few cases, the same message can have hits on more than one of these 
shortners

---

I SEARCHED EACH HAM AND SPAM CORPUS FOR MANY OF HUNDREDS OF URL SHORTNERS

HERE ARE THE RESULTS:

---
STATS FROM SPAM:
286 total spams blocked that had a shortner, out of hundreds of URL 
shortners I had searched on
(<10 that *MIGHT* be FPs - they were definitely questionable at best - 
btw, zero of those

Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen

On 2/20/2018 12:21 PM, Reindl Harald wrote:

we have well working outbound spamfilters


Excellent!

but just because someone has a google-shortener within a mail says 
*nothing at all* - frankly i even got a week ago a mail from my boss 
where the google-shortener was used for a only internal reachable 
server with a long list of params in the url
and hence the google-shortener don't say anything 



This is a very abused loophole that says MUCH in certain contexts. And 
I've carefully constructed these change at invaluement to be extremely 
unlikely to impact those who are using "goo .gl" for legitimate purposes 
and are not using it to cloak their domain in messages that are UBE or 
otherwise not desired by the recipients. But, in contrast, marketers 
and/or ESPs who start doing this routingly, as a purposeful regular 
practice, and who don't have some kind of real and specific purpose such 
as what that you described, are essentially giving DNSBLs and spam 
filters "the middle finger". So I'm giving it back. ANYTHING that 
facilitates anonymizing identity is VERY BAD for email. Facilitating 
anonymizing identity causes more spam to be delivered and punishes good 
senders when bad senders get away with that. Methods that facilitates 
anonymizing identity for email is not something that anyone should 
defend or celebrate - even if anonymizing identity wasn't the original 
intended purpose. I understand your very legitimate concern that this 
crackdown might lead to collateral damage. That is admirable. But 
acceptance of a new and pervasive situation in email that anonymizes 
identity is a HUGE step backwards... like going back to the mid 2000s, 
or something. So some "push back" measures are exceedingly warranted.


--
Rob McEwen
https://www.invaluement.com
+1 (478) 475-9032




Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen

On 2/20/2018 11:45 AM, Rob McEwen wrote:
And we ALL have to constantly shift our tactics to deal with emerging 
realities like this one - or risk getting left behind by our 
competitors who do keep up.



ALSO - Likewise, it was very frustrating that I had to spend hours late 
last night making adjustments to my spam filter to be able to block more 
of these egregious spams that are often so difficult for a spam filter 
to block - particularly since extra careful measures are needed to keep 
such adjustments from blocking legitimate messages. (I've been trying to 
get to that for weeks, as this problem has been festering for some time 
- but I just recently recovered from the flu.)


So nobody should cry on my shoulder about the difficulty of ESPs doing 
such abuse monitoring/prevention. Any such person doing so is likely 
underestimating the current size and frustration that spam filtering 
admins are having with this problem!


--
Rob McEwen
https://www.invaluement.com
 



Re: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen

On 2/20/2018 10:57 AM, Reindl Harald wrote:
and how do you imagine that i prevent paying customers to use whatever 
url-shortener?


Perhaps use the SAME methods that an ESP would use to prevent a customer 
from sending an egregious phish (or terminate their account for sending 
a phish). Of course, I also recognize that an egregious phish is much 
worse. But my point is that such abuse monitoring and prevention is a 
real thing for ESPs! Yes, some ESPs are more sophisticated than others, 
where they do a better job at this than others. For example, I've 
received two egregious phishes to my own email address, from MailGun IP 
space, within the past several months. I alerted them in both instances 
and hopefully they are improving their system? In contrast, I don't 
think I've ever seen such a phish from Exact Target, from example. That 
isn't by accident! Some do a better job of this than others. And even 
though no ESP can be perfect - that doesn't mean they can't improve. And 
we ALL have to constantly shift our tactics to deal with emerging 
realities like this one - or risk getting left behind by our competitors 
who do keep up.


Also, getting ESPs to pass this message on to their clients, even if 
just adding this to their instructions for clients, even if just as a 
"best practices" warning... might also go a long way.


when you start list to many legit servers because of that you RBL will 
be no longer useable for responsible admins which primary job is 
receiove and deliver email and not to reject it 



I'm extremely confident that this won't happen. Most likely, a few 
marginal ESPs and marketers will get blacklisted who were previously 
just barely avoiding detection. Also, we OFTEN get outliers (such as an 
occasional VERY bad spam that came from a normally VERY good sender), 
and "decoys", too! In those cases, if those messages had led to an 
automatic blacklisting, and we didn't FIRST check those domains and IPs 
against our very sophisticated "false positive prevention filter" - then 
what you described - would have happened a long time ago already. But, 
instead, invaluement's reputation for low False Positives speaks for 
itself. Given what I know about how invaluement works "under the hood", 
I can say with confidence that it is practically impossible for this 
change to put a dent in our hard-earned low-FP reputation. But this 
COULD cause problems for some already dark-gray-hat ESPs who let this 
practice run rampant.


--
Rob McEwen
https://www.invaluement.com
 



The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

2018-02-20 Thread Rob McEwen

RE: The "goo.gl" shortner is OUT OF CONTROL (+ invaluement's response)

WARNING FOR ESPs AND MARKETERS: Google's "goo.gl" shortner is OUT OF 
CONTROL.


Spammers are starting to use this to evade spam filters, and Google 
isn't keeping up with the abuse, nor shutting these down fast enough. 
Along with blackhat spammers, we're seeing evidence that many 
gray-ish-hat spammers are jumping on this bandwagon, in the hopes that 
they will get more mail delivered if they can avoid using their domain 
in the clickable links. Keep in mind that, if a marketer is doing things 
the right way, they should have no need to obfuscate their own domain 
name. They should instead proudly use it and not feel the need to hide 
behind Google's shortner. Yes, there are many legitimate uses of 
Google's shortner, too. However, we are now at a point where a VERY 
large % (a majority?) of uses of these headed to a typical user's 
mailbox are egregious spams, and a significant additional portion are 
likely-spams. THEREFORE: If you like having NOT-blacklisted IPs, be 
advised that the invaluement anti-spam DNSBL system is now adding "bad" 
points to the scoring of all messages that use the "goo.gl" shortner, 
and we're amplifying other "bad" points. (We're also doing various 
sophisticated things to minimize potential resulting FPs, too. But this 
will still put MANY marginal IPs and domains into our blacklists that 
normally might have barely avoided an invaluement blacklisting!)


YOU HAVE BEEN WARNED! DON'T ALLOW YOUR CLIENTS TO DO THIS!

ALSO: We're very seriously evaluating the option of converting each 
shortner to the URL it redirects to - and then potentially starting to 
add those domains or IPs within those URLs to our ivmURI domain/URI 
blacklist. This might not cause other such messages to get blocked, but 
it will have other negative repercussions for other uses of that domain.


--
Rob McEwen
https://www.invaluement.com





Re: Blacklist for reply-to?

2018-02-18 Thread Rob McEwen

On 2/18/2018 3:06 PM, Kenneth Porter wrote:

Is there a blacklist for domains in the reply-to header?
I've noticed a lot of spam with no URL and mutating From but the 
reply-to domain is always aliyun dot com. I want to add a site-wide 
blacklist for that. 



http://msbl.org

(I'm not associated with this. Also, it is very high quality and 
well-run! It should at least make a noticeable improvement, even if it 
doesn't catch all of them.)


--
Rob McEwen
https://www.invaluement.com
 



Re: smtp.centurylink.net 206.152.134.66

2018-02-11 Thread Rob McEwen

On 2/11/2018 2:37 PM, David Jones wrote:
This mail server has legit email for centurylink.net and 
embarqmail.com plus a lot of other spam coming out of it.
It's listed on a number of RBLs making this very hard to allow ham 
through and block spam.

http://multirbl.valli.org/lookup/206.152.134.66.html

https://pastebin.com/YidWCqp8


I've downgraded the whitelisting entry for this IP at invaluement. It 
still won't get blacklisted due to the large amount of collateral damage 
that such a listing would cause. (And others lists having this 
blacklisted is probably a GOOD thing! I'm not disputing their decision 
for their list. Different lists serve different purposes, etc.) But with 
this downgrade at invaluement, future spam that comes from this IP will  
be examined with greater scrutiny by invaluement, in order to possibly 
blacklist other domains and IPs related to the spam.


Also, the spam sample shows a Google shortner being used as the payload 
link. I've seen many of those lately - and I think Google needs to work 
on improving their ability to prevent these, or at least get the 
shortner terminated faster. At the moment, this one is still "live". I 
reported this particular one as spam to their shortner abuse form. So, 
it will be interesting to see how long it persists from this point forward?


btw - if anyone ever wants to learn more about one of these google 
shortners without actually visiting the link (which can be dangerous... 
for example, some of the more malicious links arrive at a page that 
tries to install a virus), add ".info" to the end of the google shortner 
URL and you can then see more info about the shortner, including its 
intended destination. For example, for this one:


https://goo.gl/s7XxhD.info

--
Rob McEwen
https://www.invaluement.com




potential new SA feature: Direct DNS Querying Per DNSBL Zone

2017-11-15 Thread Rob McEwen
 
this completed - that would be greatly appreciated!


NOTE: Once this feature is completed and bug free, it will then require 
the approval of the Apache Software Foundation (or something like that - 
I don't really understand their inner workings very well) - before it 
can be officially adopted. However, already, it has significant support 
from multiple important people! Therefore, I have high hopes that it 
will eventually earn this adoption, but that is not a guarantee and I 
have no control or say in that decision. But it makes the most sense to 
FIRST make this potential SA addition bug-free and outstanding - then 
cross that other bridge when we get there.


Thanks to anyone who is willing/able to help with this. Please let me 
know if you have any questions. I'm also considering putting some kind 
of modest "bounty" on this - to fund anyone's efforts that either fix 
the remaining bugs, or at least make significant and measurable progress 
to that end - send me a private message off-list if that interests you! 
(I would do this myself, but Perl "looks like Greek" to me!)


https://bz.apache.org/SpamAssassin/show_bug.cgi?id=7477

--
Rob McEwen
https://www.invaluement.com





Re: Weird new malware

2017-11-08 Thread Rob McEwen

This seems to be catching most of them:

Subject: Invoice [A-Z]{2,3}\d{7}\b
...but it might need to be combined with other things to ensure no false 
positives, since there would be a rare legit message that would hit on this?

--Rob McEwen

On 11/8/2017 10:45 AM, Dianne Skoll wrote:

Hi,

Heads-up: We're seeing weird new malware with a subject that looks like

Invoice XXX

where XXX is two or three random upper-case letters and n is a series
of digits.  What's weird is that the Content-Type: header looks like this:

Content-Type: multXXXart/mixed

where the XXX is the same as in the subect.  That is, a message
with subject "Invoice UUI8187685" has Content-Type "multUUIart/mixed".  This
is fooling our MIME parser because it doesn't see the container as a
multipart.  Does any client software?

Anyway, might want to make rules for this.

Regards,

Dianne.



--
Rob McEwen



Re: Blocking senders that are whitelisted

2017-10-04 Thread Rob McEwen

Alex,

(if you have the time...) You should challenge this sender to provide 
evidence that your user signed up for their messages. Tell them that it 
isn't good enough to receive an IP address and date/time-stamp. You want 
to know what action your user took to get on their distribution list. 
(then possibly share that information here and with Mail Chimp)


It could very well be that your user signed up for their services in 
some way - but either forgot - or the messages got rebranded in such a 
way that your user didn't recognize it? But it is also possible that 
they were added via a purchased list or something bad like that.


Rob McEwen
invaluement.com

On 10/4/2017 11:23 AM, Alex wrote:

Hi, we have a user complaining about receiving email from a solar
panel company and want us to block it. The problem is that it
originates from mailchimp, which is whitelisted.

It's my belief that mailchimp is safe to whitelist (mcsv.net).
However, what happens when an email is received that needs to be
blocked? Do you just report it?

Chances are probably good that I could click the "unsubscribe" link
and just unsubscribe the user, and will probably do that, but I'm
looking for a more general solution.

I'll also report the sender to mailchimp, but the email looks like
they're following all the rules. The sender is
exped...@iowawindandsolar.com.

Should I block the From address in postfix?

Is it possible to blacklist the From when the sender is whitelisted?



--
Rob McEwen
http://www.invaluement.com



Re: ramsonware URI list

2017-07-15 Thread Rob McEwen

On 7/15/2017 2:13 PM, David B Funk wrote:

How quickly do stale entries get removed from it?


I randomly sorted this list, then I tried visiting 10 randomly selected 
links. I know that isn't a very large sample size, but it is a strong 
indicator since they were purely randomly chosen. 9 of the 10 links had 
already been taken down. So there might be much stale data in that list?


I also extracted out the host names, deleted duplicates, randomly sorted 
those, then ran checks of 500 randomly selected host names against 
SURBL, URIBL, DBL, and ivmURI. The number of hits on all 4 lists of 
shockingly low. But I think that probably has more to do with stale data 
on this URL list (and this is really a URL list, not a URI list), rather 
than with lack of effectiveness of these other domain/URI blacklists.


Still, there can be situations where a URI list won't list such a host 
name due to too much collateral damage - but yet where a URL list that 
specifically lists the entire URL - can still be effective.


Because such a URL list would be LESS efficient (due to being 
rules-based), it would be preferable that such a list would have much 
less stale data - and perhaps would focus on the stuff that isn't found 
on any (or very many) of the 4 major URI lists I mentioned, so as to 
keep the data small and focused, for maximum processing efficiency.


--
Rob McEwen
http://www.invaluement.com


Re: URIBL_BLOCKED on 2 Fedora 25 servers with working dnsmasq, w/ NetworkManager service

2017-05-18 Thread Rob McEwen

On 5/18/2017 5:46 PM, David Jones wrote:

it should be pretty clear now to not use a forwarding DNS server locally and
do not point the server to another DNS server in /etc/resolv.conf.


Thanks David!

Some may be interested to know at least 15% of my entire labor 
"overhead" for running invaluement - involves playing "whack a mole" (so 
to speak) with both testers and existing subscribers - whose DNS 
settings CONSTANTLY revert back to sending direct queries to invaluement 
via Google and/or OpenDNS - which are then blocked - even as the 
instructions were extremely clear about how/why not to do it that way.


In many cases, they explain to me that their settings got 
auto-overwritten by their hoster - who just HAD to switch their 
resolv.conf file back to 8.8.8.8


In some rare worst case scenarios - I have to "fire the customer", due 
to many repeated incidents where the labor involved in constantly 
babysitting their settings - was no longer worth their subscription payment.


And unfortunately there is just basically a very sizable portion of IT 
professionals in the entire world... probably hundreds of thousands of 
IT people - who have been convinced that pointing all DNS to 8.8.8.8 is 
standard operating procedure that they think is always the best way.


For me, it feels like annoying busy work. Imagine that for at least one 
hour out of your day - you have to stop what you're doing and dig a hole 
in your back yard - and then fill it back in.


So I'm grateful every time I see thread like this that pushes back 
against that, and encourages others to run industry standard 
non-forwarding caching DNS servers.


THANKS!

--
Rob McEwen
http://www.invaluement.com




Razor FP on simple http link (by itself)

2017-05-05 Thread Rob McEwen
I use SA as a "helper app" within my custom written spam filter. So I'll 
get SA give me an opinion about certain marginal messages, and then my 
spam filter factors the SA score into my spam filter's scoring.


Recently, a prominent law firm for whom I host mail - was complaining 
about FPs where messages from a prominent real estate company were not 
making it to them. Interestingly, their messages kept hitting RAZOR, 
where SA was giving the following response:


1.7 RAZOR2_CHECK   Listed in Razor2 (http://razor.sf.net/)
0.4 RAZOR2_CF_RANGE_51_100 Razor2 gives confidence level above 50%
   [cf: 100]
2.4 RAZOR2_CF_RANGE_E8_51_100 Razor2 gives engine 8 confidence level
   above 50%
   [cf: 100]

In testing, I narrowed it all the way down to simply the following 
(alone!) hitting on razor:


either
http://www.example.com
or
http://example.com

(except with the sender's domain, of course)

...either one was triggering this razor score. I even put that as the 
ONLY body text of another message (so a totally different header) - and 
it still triggered. But either variation WITHOUT the "http://; part did 
not trigger.


Interesting... this domain name happens to resolve to an IP that is 
currently blacklisted on Zen. (I know, that is really really bad!) 
Unfortunately, that confuses issues!


Does RAZOR extract domains from links and checks them against a bad 
domain database... sort of how SURBL works... and/or check the IP that 
they resolve to? (I don't think so, but now I have to ask just to be sure!)


If not... this seems to go beyond checksum-checking of parts of a 
message - this seems much more surgical/specific than that.


Don't get me wrong... I'm a big fan of razor and of other 
checksum-technologies. But I'm sort of shaken by this because I always 
thought a FP for razor would be much more difficult due to larger 
portions of a message having to match a checksum match in order to have 
a hit. (sort of like a larger "fingerprint" that is not easily 
duplicated in another innocent message, allegedly making FPs practically 
impossible)


While this kind of more surgical strike can be beneficial in blocking 
more spam - it seems like it changes the paradigm of what I 
(mistakenly?) thought to be RAZOR's potential for collateral damage.


Is this "extra curricular activity"? or did I misunderstand RAZOR's 
checksum technique?


--
Rob McEwen


Re: Outgoing email without DMARC

2017-05-02 Thread Rob McEwen

On 5/1/2017 10:30 PM, Marc Perkel wrote:

Might be slightly off topic but I've been running into more delivery
problems with outgoing email because I don't use DMARC. I don't know a
lot about it but is there some simple way I can get around this? Kind of
a pain in the rear.


Marc,

This probably has more to do with DKIM than DMARC?

Either way... you're not willing to jump (or haven't yet jumped) though 
the hoops that the largest ISPs/hosters want us all to jump through... 
meanwhile... so many of them (and for many many years) have sent such 
high volumes AND high percentages of outbound spam to all of our SMTPs - 
to such an extent that you and I would be out of business if our SMTP 
outbound traffic did that for just one week.


I sort of wish they (or many of them... "if the shoe fits...") would 
clean up their own act FIRST - get the basics done FIRST - before 
imposing new standards on the rest of us.


I'm in the same boat - I'm now having to set aside dozens of hours to 
get all various domains updated to DKIM so that they'll have more 
success sending to a certain large/famous hoster - who has sent my 
server a shitload of spam over the past several years (not just 
volume-wise - but percentage-wise... I'd be run out of town if I did that)


--
Rob McEwen




Re: Fastest listing RBL ?

2017-02-15 Thread Rob McEwen

On 2/14/2017 11:04 PM, Ian Zimmerman wrote:

Given a piece of horrible spam, on which RBL is the sending IP address
likely to appear first?

I want to rationally decide which RBL/s to consult at SMTP time.  Afraid
to use all of them, not just due to false positives, but also due to
negative caching in DNS, which could affect the result when the spam is
seen by SA a bit later.


Ian,

(if possible) Start monitoring the sending IPs of barely missed incoming 
spam, in real time, as they come in. As soon as possible after the spam 
comes in, check that IP on multirbl.valli.org (or mxtoolbox), and see 
which RBLs are listing that IP (recognizing that SOME of those are going 
to be situations where the IP wasn't blacklisted until seconds after the 
spam was received - but many were listed before then)


Throw out the examples where the IP has a good reputation in places like 
SenderScore or SenderBase... or where the IP is an MTA of a large/famous 
hoster/ISP (but take note of which lists were listing those) - However, 
SOME of those listings with decent scores at SenderScore or SenderBase 
are appropriately blacklisted, such as a small sender with a massive 
security hole where they are suddenly spewing out a massive amount of 
spam. But if you notice an RBL hitting on MANY like this - they might be 
too aggressive and even worthless - or perhaps more appropriate for low 
scoring. There are a few lists which are VERY fast-reacting to new 
spams, but have horrific expiration polices and/or are poorly managed - 
and would produce many FPs if used in production. But there are some 
other lists which are just a little too aggressive for outright 
blocking, but are very useful for scoring a few points (or less).


You should then notice about only about ~5-8 lists which rise to the top 
- in the checking I described, they seem to have very few FPs (you won't 
know for sure until you test them in production) - and they are very 
fast reacting. Still, SOME of these are just a little too aggressive for 
outright blocking (or scoring at/above threshold) - but are great for 
high scoring. But you may not be able to accurately measure their FP 
level until you've used them in production for some weeks.


A smaller amount of these RBLs (that rose to the top)... in addition to 
being fast-reacting... block much unique spam missed by the other lists 
AND have few enough FPs for you to probably feel comfortable outright 
blocking (or scoring at/above threshold). You might find ~3-5 such 
lists, including zen.spamhaus.org in that elite group.


--
Rob McEwen




Re: The nice thing about standards (was Re: Legit Yahoo mail servers list)

2017-01-31 Thread Rob McEwen

On 2/1/2017 12:56 AM, Dave Warren wrote:

They publish SPF records and DKIM sign everything for competent SMTP
receivers to handle in real-time, AND they publish a HTML version for
humans, and yet someone still finds a reason to complain?


Dave,

After the initial question was raised, it took about 11 posts and almost 
24 hours for someone to notice the discussion who happened to know about 
the "HTML version for humans" and mention that. During those 11 posts, a 
well-respected and knowledgeable person was actually defending Yahoo for 
NOT having such a page, which gave the impression that such didn't 
exist. (certainly, that was a head-fake that I fell for, even if such 
was very innocent)


So I think there is a strong argument that the existence of this page 
page isn't exactly common knowledge. Archive.org suggests that this page 
has only existed for a couple of years. I've been looking for it 
(occasionally) for the past 10 years - so I think all my memories of 
past discussions in past years about such a page not existing - were 
probably accurate. By the time this page existed, I had given up on 
finding it. (not that I spend every waking hour looking for it - I think 
I probably looked for it about once every year or two - for some time - 
and the need for this isn't so great with other senders - because few 
senders [even large ones] have such a MASSIVE amount of sending IPs that 
are so particularly hard to find)


Regarding your references about such a page not being needed - all I'm 
going to say is that some systems benefit from having large IP ranges 
preemptively whitelisted for the sake of efficiency. There are scenarios 
in certain very high volume systems where this enables the processing of 
messages at order of magnitudes faster rates than if SPF and DKIM and 
FCrDNS-confirmation had to be checked on every sending IP. MUCH of that 
relies on the response times of 3rd party servers - which (even at 
best!) is order of magnitudes slower than a local rbldnsd query  - or 
than an optimized binary search of an in-memory array - which is even 
faster than rbldnsd or even a high-end in-memory database. Sometimes, 
such 3rd party servers can "freeze up" in their responses, or rate limit 
queries - or firewall such lookups for what is perceived as abuse - 
causing further complications. Caching only does so much to prevent this!


That kind of need for speed is the world in which I live. At 
invaluement, I'm processing dozens of spams per second - and since much 
of these are ones where the "low-hanging fruit" - such as ALREADY 
heavily blacklisted botnet-sent spams are ALREADY filtered out before 
they get to my system - that means that the processing resources per 
spam is already much higher for my system than that of a typical ISP or 
hoster's natural incoming spam. (I process a higher concentration of the 
more sneaky spams and the newer emitters)


With this in mind... if I deleted my IP whitelist, and had to rely on 
SPF and DKIM and FCrDNS-verification for EVERY message, my queues would 
back up considerably - and a lot of worthy blacklistings of IPs and 
domains from new incoming spams would get considerably delayed. (again, 
inevitably - at this volume - issues come up where such 
queries/verification suddenly "freeze up" or get rate limited, 
firewalled, etc)


And I think my need for efficiency is probably not much different than 
some very large hosters and ISPs - who process mail for millions of users?


And I think we've already established that there is no possible way to 
generate "on demand" and remotely efficiently the information on that 
HTML page just via Yahoo's SPF records.


iow - maybe you should have a little more respect and try to be a little 
less snarky in the future - when you don't necessarily know/understand 
others' situation/requirements that may be a little different than your 
particular situation/requirements.


--
Rob McEwen




Re: Legit Yahoo mail servers list

2017-01-30 Thread Rob McEwen

On 1/30/2017 8:54 AM, Matus UHLAR - fantomas wrote:

they do and it has been mentioned:
https://help.yahoo.com/kb/SLN23997.html


I wasn't aware of this page. If it was mentioned before in this thread, 
I missed it. Thanks!


--
Rob McEwen




Re: Legit Yahoo mail servers list

2017-01-29 Thread Rob McEwen

On 1/29/2017 7:42 PM, Dianne Skoll wrote:

On Sat, 28 Jan 2017 16:33:24 +
David Jones <djo...@ena.com> wrote:


Read back through this thread.  I never said their SPF record is
invalid. All I said is their SPF record is not common and it makes it
very hard for anyone to know what the official Yahoo outbound mail
servers are.


Why is that important?  Can't you just whitelist the domain yahoo.com if
and only if it hits SPF "pass"?


We have to work very hard to get our MTAs to whitelist
them.  It's in their own best interest to make this information
easily available to the Internet since so much spam comes out of
their platform.


Then why would you whitelist them?


They are too large to not whitelist.


Nobody is too large to not whitelist.  They're obviously too large to
block, but you'd be foolish to accept any and all mail from a Yahoo
server unless you like an awful lot of spam.

Regards,

Dianne.




Dianne,

I can't speak for David, but most or all of your answers don't apply to 
my own anti-spam blacklist's attempt to try to avoid blacklisting Yahoo 
IPs that are both known for sending much spam, but which also would have 
a very high rate of collateral damage if blacklisted. (recognizing that 
some very good DNSBLs, which are more aggressive, are more willing to 
blacklist Yahoo IPs, and that isn't always a bad thing)


...and/or your answer requires more on-going receiver-side resources.

Interestingly, many senders would crawl over broken glass if necessary 
to provide me their IPs, if said I was seeking those for my whitelist.


Also, when David said "whitelist", I can take an educated guess that he 
isn't allowing Yahoo-sent messages free unfiltered access to the inbox - 
he is probably just trying to avoid DNSBL checking of those particular 
IPs - but then he'll probably STILL do other content filtering of those 
messages. That would be my educated guess. And this would be a SMART 
strategy.


Personally, when I get messages from Yahoo into my hosting business - I 
have the IPs generally not checked - since I already have most Yahoo IPs 
whitelisted - then I only content-check the messages - BUT... next I 
AMPLY any content scoring of such messages since these came from Yahoo 
and are more likely to be spam - that is, if the sender isn't already in 
a carefully cultivated exception list of known good Yahoo senders 
(specific to my mail hosting user base) - I do this for all freemail 
senders known to send a high volume of spam.


I know you mentioned that Yahoo may want to have the flexibility to 
change their IPs. But instead of providing a list, they could also 
provide a link to a web page listing the IPs (like what Comcast does) - 
and then just update that web page whenever their IPs change. This isn't 
rocket science.


As it stands, it is mind boggling just how many Yahoo ranges of sending 
IPs there are worldwide. Over the years, I've added 53 yahoo entries to 
my whitelist. Besides the hundreds and hundreds of /24s ranges in there 
(many are multiple consecutive /24s, showing up as just one line of 
those 53 entries), there are also several /16s, too. It would be nice to 
be able to compare that to Yahoo's current list active sending IPs (if 
such were available?), so that I could EFFICIENTLY update/prune that 
part of my whitelist.


And I strongly suspect that iterating though the millions of IPs to 
check FCrDNS would take a very, very long time - and might get such 
probing IPs blacklisted for abuse/intrusion-protection?


--
Rob McEwen




Re: Legit Yahoo mail servers list

2017-01-27 Thread Rob McEwen
While I have Yahoo sending IPs extensively covered in my whitelist, I've 
been trying to get their complete official list of sending IPs for years.


I'm amazed that Yahoo doesn't participate in these conversations - or 
that nobody ever says, "I'll ask my colleague over at Yahoo"


seems very odd...

--
Rob McEwen




Re: How to create a URIBL

2016-10-19 Thread Rob McEwen

On 10/18/2016 9:09 PM, Alex wrote:

How do you then enter ranges? For example, one of the rbldnsd zone
examples I've seen have entries such as:
1.168.160.0-255
That does not look to be in reverse order, as the host octet is still last.


while there may be a more complicated and unusual answer for this.. the 
short answer is... you don't, and you shouldn't have to.


(1) IPs at the base of clickable links inside the body of the message in 
spams... is still a little rare... comprising roughly 2% of all such 
listings.


(2) This means that (a) those IPs aren't taking up a lot of space in the 
dnset files, when compared to the domains and host names there, and (b) 
of that ~2% of IPs, extremely few of those are even in the same /24 
block - so you don't get much mileage out of trying to list ranges


having said that... sending-IP lists that use ipset DO have the 
functionality that you desire. ipset actually has quite a number of 
acceptable formats to list blocks or ranges of IPs.


iptset... not so much. iptset is built for EXTRA speed and EXTRA 
low-memory usage, but isn't as flexible and generally requires one 
single IP per line.


Based on your question, it could be that you're trying to merge your 
sending IP blacklist, with your URI/domain blacklists... all into one 
single dnset rbldnsd file? if so, that is NOT recommended. It causes 
problems and removes some of rbldnsd best features/strengths.



Your service is great, btw.


Thanks. Please send me a note off-list as you how/why you think that. 
I'm not looking for praise... just curious if you're one of my clients 
(such as at your dayjob?) or if we've crossed paths somewhere and I 
forgot about it?... or if you have ever testing invaluement? etc (though 
I know you're a frequent SA discussion participant)



--
Rob McEwen
http://www.invaluement.com
+1 (478) 475-9032




Re: How to create a URIBL

2016-10-19 Thread Rob McEwen

On 10/19/2016 3:51 AM, Matus UHLAR - fantomas wrote:

are you REALLY sure the IP has to be reversed?
rbldns parses IP and reverses them by itself, if used in ip4* dataset.
When used in dnset, it should not be reversed.


Your most valid points do not apply to "dnset". they apply to ip4tset 
and ip4set for sending-IP blacklists.


Let me explain... but before I explain, let me say that I'm not arguing 
for any of this. These standards were put in place long before my time 
(and are followed by SURBL and URIBL, too). Or, at least I didn't set 
these standards. I MIGHT have been involved in some of the discussions 
about this circa 2004, in internal discussions at SURBL - and in SA 
discussions - but I think this was all set just a little before my time 
in those forums.


So basically, if you look at the anatomy of a domain name... from left 
to right, you get into a higher hierarchy.


So in "foo.example.com"

"foo" is drilling into detail. while "example.com" is the bigger 
picture. And then ".com" is an even bigger picture! In a domain, as you 
get FURTHER to the right, you go to a HIGHER hierarchy or level.


But IPs are the opposite. For an IPv4 IP, the leftmost number is the 
highest in the hierarchy, and you drill down into more detail as you 
move to the right.


For this reason, it was decided a long time ago... that for URI DNSBL 
blacklists that use "dnset", the IP should be reversed in the source file.


Therefore, in the data file, the test point IP:

127.0.0.1

shows up as

1.0.0.127

And then when the client queries that IP, the query is formatted as follows:

1.0.0.127.example.com

(where example.com is the URI blacklist's host name)

And, likewise, ALL of the major anti-spam software, (such as 
SpamAssassin), automatically reverses the IP when that (forward-ordered) 
IP is extracted from a base of a URL found in the body of a spam, and 
then this is appended to the beginning of a URI blacklist's hostname, 
for checking against a URIBL blacklists (such as SURBL, URIBL, or my own 
ivmURI list)


This decision to do it this way PROBABLY had something to do with trying 
to get rbldnsd engine to NOT have to internally treat IPs and 
domains/host-names differently. otherwise, it would have had to "know" 
to reverse IPs, but yet know to NOT reverse domains or host names. (and 
who knows what TLDs could be coming up in the future?)


In contrast, IPs found in sending IP data files (for ip4tset and ip4set) 
don't have this inconsistency problem. So it make sense to just leave 
them in forward-order, for EASY readability... and then just allow 
rbldnsd to reverse order them on-the-fly. (thank God - I'd go nuts if my 
ip4tset and ip4set were all in reverse order! meanwhile, IPs in URIBL 
data files are usually a TINY percentage of the listings!)


--

Having said all of that, for regular sending0IP blacklists, (just as you 
said) the IP is NOT in reverse order in the file. But rbldnsd "knows" to 
reverse order it in memory, before it is compared to the reverse-ordered 
query that comes in from the client.


So you're correct when you say, "rbldns parses IP and reverses them by 
itself" ... but that only applies to sending-IP blacklists, set up with 
ip4tset and ip4set in rbldnsd.


As shown, dnset operates differently for IP addresses found in URIBL 
blacklists.


----------

This was a trip down memory lane for me.

--
Rob McEwen
invaluement


Re: How to create a URIBL

2016-10-18 Thread Rob McEwen

Alex,

here are some suggestions:

In your rbldnsd-formatted file, put a dot at the beginning, which serves 
as a wildcard.


So your three examples:

109 .73 .134 .241
51steel1 .org
amessofblues1 .com

(I added spaces here to evade spam filtering, but those spaces shouldn't 
actually be there)


would like like this:

.241 .134 .73 .109
.51steel1 .org
.amessofblues1 .com

(again, the extra spaces shouldn't be there)

NOTICE 2 things:

(1) The extra dot at the beginning
-and-
(2) the fact that the IP is in reverse order. The great part about 
rbldnsd is that a lookup on either


example.com
OR
www.example.com
OR
foo.bar.foo.example.com

ALL of those will get a "hit" when the rbldnsd file has

.example.com



When it comes to formatting the rbldnsd-formatted file, in addition to 
my suggestions above, it comes down to a choice... make it a simply list 
of the domains and (reverse-ordered) IPs? Or provide more information 
for each individual IP, such as a custom text response, as you did here:


foo.example.com:127.0.0.2:Blocked System

in my experience, I haven't been able to get this to work unless I put a 
space just before the first colon, as follows


foo.example.com :127.0.0.2:Blocked System

But sometimes you don't need that and can simply use just the domain or 
IP on each line, since much of that can be accomplished with a single 
line at/near the top of the file, such as this one that I use for the 
invaluement URI list:


:127.0.0.2:Blocked by ivmURI - see http://www.invaluement.com/lookup/?item=$

...which then causes all following lines of just domains and IPs... to 
use this line above as if it were on every single line. - and the "$" 
causes the actual listed item to show up in the SMTP text message. That 
"$" feature can be very informative and helpful!


of course, the most difficult part is not collecting spammy IPs and 
domains... that part is easy. The most difficult part is knowing when 
NOT to blacklist a domain--which would be a decoy domain found in a 
spam, that wasn't the actual "payload" for the spam and is instead an 
innocent bystander's domain -- and/or generally keeping FPs super low. 
THAT is the hard part.


There are other issues as to WHERE to divide the domain.

For example, if you listed

.foo.bar.foo.bar.foo.bar.foo.bar.example.com

... but foo.bar.foo.bar.foo.bar.foo.bar. was just decoy material added 
by the spammer... then...


foo.bar.example.com comes in and guess what? your lookup fails to find 
it. Yet all such variations would be listed if you had simply blacklisted:


.example.com
(again, with the dot in front)

But try this and blacklist:

.blogspot.com

...and trigger massive FPs... when you should have listed:

.somehorrificspammerfromhell.blogspot.com

so that either

www.somehorrificspammerfromhell.blogspot.com
OR
somehorrificspammerfromhell.blogspot.com
foo.bar.foo.bar.somehorrificspammerfromhell.blogspot.com

would ALL return listing, but

blogspot.com

...wouldn't.

So it also takes some work determining those boundaries. Some of those 
are simple domains... while others like blogspot.com or wordpress.com, 
are more "artificial" (but still critically important).



--
Rob McEwen
invaluement.com



Re: RCVD_IN_SORBS_SPAM and google IPs

2016-09-12 Thread Rob McEwen

On 9/12/2016 9:37 AM, David Jones wrote:

The majority of the junk can be blocked with zen.spamhaus.org and
sip.invaluement.com RBLs.  Every small mail filtering platform should
use zen.spamhaus.org as long as they are under the free usage limit.
The sip.invaluement.com is a private RBL but very reasonably priced
and is a great complement to zen.spamhaus.org.  The major senders
should not be listed in these 2 major RBLs so they fit right in with the
6 items above.


David,

Thanks so much for the recommendation!

But I just to make sure that everyone knows that throwing 
sip.invaluement.com into their spam filter is not going to help anyone 
as (1) it require a subscription to work -AND- (2) we now use DIFFERENT 
host names. In fact, eventually, no invaluement users will be using that 
host name (they'll eventually all be ported over to other alternate host 
names)... and then sip.invaluement.com will be wildcarded to "list the 
world"... eventually... (I'll make plenty of announcements before that 
time comes). Nevertheless, there will be an unfortunately group of users 
who threw sip.invaluement.com in their filter (unlicensed)... and will 
leave it there even though they've never gotten a single "hit" from 
their mis-configuration, and then they'll have a very bad day when that 
time comes.


But, again, thanks for the mention! Perhaps, next time just say 
"invaluement".

--
Rob McEwen
invaluement.com





Re: spamassassin and caching nameservers

2016-08-22 Thread Rob McEwen

On 8/22/2016 9:15 PM, Alex wrote:

Has anyone configured it as a local caching nameserver, and if so,
could you share your config?


Correct me if I'm wrong... but...

I'm almost positive that rbldnsd acts ONLY as an authoritative name 
server, and not ever as a caching name server. I don't think there is 
functionality to either fetch root hints or to do catch-all forwarding 
to an upstream DNS server for just any host names. Instead, it only 
serves up the zones that it is specifically told to serve at startup, 
using the physical source data files to which those zones point.


It was designed from the ground up only to serve as a dumbed down 
locally hosted DNS, only for serving DNSBLs where the data files are 
found locally. It makes up for the lack of more extensive DNS features 
with blazing speed and very low memory overhead.


--
Rob McEwen



Re: Spoofed Domain

2016-08-09 Thread Rob McEwen

On 8/9/2016 5:56 PM, Anthony Hoppe wrote:

Here are the headers as an example:
http://pastebin.com/bnU0npLR
This particular email has a macro-enabled Word document attached, but I
don't want to assume this will be the case every time.
Any tips/tricks/suggestions would be greatly appreciated!


I think there is a trend now... towards blocking ALL .docm files (if 
not, there should be!). I think it is EXTREMELY rare for normal human 
beings to send Word documents in that particularly dangerous format. 
Most would be send in .doc or .docx format.


I'm not sure if there is already a SA rule for scoring against .docm 
files attachments? Perhaps someone else could help you with that.


--
Rob McEwen



Re: Corpus of Spam/Ham headers(Source IP) for research

2016-06-29 Thread Rob McEwen

On 6/29/2016 1:00 AM, Shivram Krishnan wrote:

Thank you so much for your views. I agree that your customers would not
like it if you share information. But Oliver suggested , I need only the
source IP addresses of the Spam and Ham emails , which can even be
anonymized in the last octet.


Unfortunately, accuracy and credibility goes down since there then isn't 
any easy way to audit or double-check the root cause of the classification.


For example, some people classify spam as "what our filter said was 
spam" and ham as "what our filter said was ham". For most well-run 
systems, that is going to be overall very accurate. But there can still 
be egregious mistakes. And assuming that the existing filter is 100% 
accurate leaves no room for improvement. It also has the unfortunate 
side effect of rubber stamping the most elusive spams, sent by the 
shrewdest of spammers, as ham.


If an anti-spam blacklist comes along that is very good at blocking 
messages that are unsolicited and not desired by end users... but sent 
by the most shrewd spammer who evade lists like SpamHaus and SURBL (at 
least for some time)... and where the collateral damage for listing such 
domains and sending IPs is non-existent... such a blacklist might STILL 
fare badly in such a rating system... which would then MISTAKENLY assume 
that such a blacklist has many False Positives.


Stats collected from user complaints about False Negatives can also be 
helpful. However, for snowshoe spam, that is often a lagging 
indicator... sometimes days behind reality--where the spammer has 
already moved to new domains/IPs--but such could help such a ratings 
system to make wise adjustments to past ham/spam stats.


Hijacked IP and domains is another sticky issue. Over the past several 
years, this has become epidemic! If the volume of legit usage is 
relatively low, and the IP or domain has been hijacked by a spammer... 
then at SOME point, an anti-spam blacklist should not be penalized for 
listing such. In fact, Spamhaus does this frequently (lists hijacked 
domains/IPs where the cost/benefit ratio for that listing is well 
justified). Some other lists also blacklist hijacked domains/IPs... but 
are often not as good at making proper cost/benefit ratio decisions... 
where they list somewhat large senders who had a somewhat small and 
short-lived spam outbreak. Finding a way to penalize or reward the lists 
that block hijacked domains/IPs that Spamhaus misses, based on whether 
they do (or don't do) a good job of making overall good decisions about 
the cost/benefit ration of a potential listing's collateral damage... is 
also tricky.


My main point is... how to reward blacklists that are more accurate, but 
without penalizing them for not being a redundant copy of Zen. It isn't 
as easy as it sounds in a ratings system. (even if real life usage of 
such by a hoster or ISP can quickly lead to fewer complains from 
customers about about FP and FNs)


--
Rob McEwen




Re: Which DNSBLs do you use?

2016-06-16 Thread Rob McEwen

On 6/16/2016 9:49 AM, Alessio Cecchi wrote:

Probably zen.spamhaus.org is the best dnsbl but is too expensive for us.
Invaluement SIP is almost comparable to Zen as performance but much less
expensive.


Thanks, Alessio, for the recommendation.

But I need to make one clarification... SIP and SIP24 should not be 
considered a replacement for ZEN because they purposely do NOT try to 
"catch every botnet" and instead focus on the more sneaky spams as well 
as new emitters. If someone tries to replace Zen with SIP and SIP24 
(combined) they would usually be very disappointed in their overall spam 
filtering, unless... as I presume to be the case for Alessio Cecchi ... 
they had other very good measures in place for blocking botnet spams? 
But the vast majority of invaluement users use SIP and SIP24 as a 
supplement to Zen, and find that Zen blocks much spam that invaluement 
misses.


Therefore, as I said, SIP and SIP24 (combined) are intended to be a 
supplement to Zen, not a replacement of Zen.


(just want to make sure this is clear!)

--
Rob McEwen
http://www.invaluement.com




Re: Spamassassin not capturing obvious Spam

2016-05-30 Thread Rob McEwen

On 5/30/2016 10:24 PM, Shivram Krishnan wrote:

I am testing spamassassin on a SPAM/HAM corpus of mails. Spamassassin is
not picking up an obvious spam like in this case
http://pastebin.com/MbNRNFWy .


Your pastebin example didn't show the "last external" sending IP. Could 
have have been there orginally, but was expunged from this sample? Could 
there have also been a link in the body of the message that was likewise 
removed?


it would be nice to be able to check those against respected low-FP DNSBLs.

Or, if the clickable link really wasn't in the original message, then 
this particular example was probably a rare malfunctioned spam that will 
be of no benefit to the spammer, and would then probably not be worth 
investigating since the spammer then has no incentive to keep sending 
these types.


--
Rob McEwen




Re: A Plan to Stop Violence on Social Media

2015-12-15 Thread Rob McEwen

On 12/15/2015 11:19 PM, Wrolf wrote:

Would it be practical to use the Spamassassin techniques of Bayesian
filtering and RBL lists to block ISIS on social media?


Bayes would have a high potential for false positives... such as 
blocking a news story about ISIS, or blocking a discussion about how to 
stop ISIS, or blocking an innocent discussion about the middle east, or 
an innocent discussion about Islam.


But ISIS isn't an nearly as much of an extremely fast moving and 
frequently morphing target as spam. Therefore, the techniques used don't 
have to have as much risk of FPs (since they wouldn't have to be as 
real-time-reacting to new scenarios) and should therefore be very 
surgically targeted towards 100% confirmed ISIS sources.


At the same time, beware of top-down solutions where governments impose 
free speech ("thought control") restrictions. Such powers could easily 
be abused in the future for nefarious purposes, such as suppressing 
criticism of the current party in power, etc.


This could be a "slippery slope".

--
Rob McEwen
+1 478-475-9032


Re: SpamAssassin Rules Regarding Abuse of New Top Level Domains

2015-10-20 Thread Rob McEwen

On 10/20/2015 12:13 PM, sha...@shanew.net wrote:

Unlike Larry (and others) I DO want to block the vast majority of the
new tlds, because we see nothing but spam from them (and my users tend
toward the more false-positives than false-negatives side of the
spectrum).  Rather than maintain a list of all the problematic tlds,
I'd rather have a blanket block rule with the ability whitelist the
handful that might be legit. 


Be careful about doing this for the long term. I think that spammer 
exploit new TLDs because they know that many anti-spam systems don't 
account for them correctly at first. (and/or maybe they are cheaper at 
first?). But in the longer term (years down the road).. they tend to 
move on to other ones, while the legit TLDs slowly increase. So this 
strategy can backfire in the long term. (but, of course, MMV... and some 
smaller hosters don't have to be as concerned about a few extra FPs)


--
Rob McEwen
+1 478-475-9032



Re: Return Path (TM) whitelists

2015-07-10 Thread Rob McEwen
Also, often, the Return Path certified sender is an ESP who sends for a 
variety of customers. There is not always an absolute guarantee that 
every one of that ESP's customer is ethical and truthful. A good ESP 
will quickly fire such any such bad apple customer... but some do a 
much better job than others. Some spend endless amounts of time telling 
blacklists, we're Return Path certified... and we had this bad 
customer... but we're working with that customer to purge their lists of 
complainers and bad addresses. (iow, help them listwash, keeping them 
on as customers)


ESPs are economically incentivized to keep marginal customers (or 
pretenders), and Return Path is economically incentivized to keep 
those grayhat-ESPs.


Yes, at the extremes, customers will be fired in both situations. But 
there is a lot of gray before those extremes trigger a firing. And there 
are many situations where those limits are pushed.


Having said that, those ESPs who choose to push those limits hurt 
themselves in the long run as their domains/IPs start getting dragged 
further and further down in various reputation and anti-spam filtering 
systems. But some of these are managed by 20-something-year old punk 
kids who haven't thought that far ahead.


I'm sure Return Path stops lots of this stuff but certainly, a 
significant amount of unsolicited messages can slip through the cracks.


Meanwhile, in contrast, DNSWL is NOT economically incentivized to go 
easy on gray senders. And some on this thread are not realizing that 
DNSWL has various LEVELS in its ratings of senders... where senders of 
BOTH legit mail and spam are marked accordingly. That way, you know to 
not outright block messages from certain mixed ham/spam sender's 
IPs... but you shouldn't treat them as fully whitelisted either. That is 
a big difference... therefore, most of the time that a virus-sent spam 
is sent from an IP in DNSWL, it is from an IP that is marked by DNSWL as 
a mixed source.


--
Rob McEwen
http://www.invaluement.com/
+1 478-475-9032



Re: Uptick in spam

2015-03-30 Thread Rob McEwen

On 3/30/2015 11:49 AM, Kris Deugau wrote:

Seconded;  this is exactly what we've been finding.  Invaluement is a
great complement to Spamhaus for a fraction of the cost.

I wouldn't put it as a front-line reject DNSBL, because some of the
things that have been listed are not what I would class, for our
customers, as spam - but those entries are distinctly greyhat at best in
a lot of cases, and some IP range operators I've flagged as list,
delist, and whitelist_from_rcvd as needed due to the mix of legitimate
small senders and spammers.


Thanks Kris for the compliment. Also, when you say mix of legitimate 
small senders ...just to clarify, I think that any further analysis 
will show that (a) MOST of these are situations where very small senders 
had massive spam-sending outbreaks due to compromised accounts, and (b) 
the listing was most often very short lived (often mere hours).


This is a balancing act... and I think invaluement strikes a great 
balance. And even in THIS particular area, I think our FP level is still 
distinctly LESS than UCEProtect, Barracuda, and SORBS (for examples). 
But if we brought that all the way to zero, MUCH spam that slips past 
Zen wouldn't be listed on invaluement anymore. (the ham/spam ratios on 
some of these compromised account situations is horrendous--they send 
out their usual 400 hams that day, along with 200,000 spams... and the 
cumulative sum total of those spams from ALL such compromised senders 
that day, represents MUCH of the spam that gets past filters due to 
piggybacking on the sender's normally good reputation)


Also, what I've found is that many medium-sized ISPs/hosters, with 10s 
of thousand of mailboxes are very comfortable with outright blocking on 
invaluement, but will only score on UCEProtect, Barracuda, and SORBS. 
Much smaller hosters will often block on all of them, because they don't 
notice those FPs as often. In fact, I see these SAME somewhat rare 
compromised-sender FPs with Zen, too. It is all about each list's 
strategies, and aggressiveness, and tolerance levels. As shown, 
invaluement is in a very strategic spot here... having much of the 
aggressiveness of these other lists, but with FP levels VERY close to 
Zen's FP levels. (and then scoring on these other lists... even 
aggressive, yet still under-threshold, scoring... will help block spams 
missed by both invaluement and spamhaus)


Also, invaluement plays close to the edge with CAN-spam and 
snowshoe spammers. So invaluement is in a little more dangerous 
territory...that it can do so and not have a lot more FPs, is not easy. 
For example, this invaluement may occasionally list the kind of pure 
ads that, upon further analysis, are arguably not technically spam, but 
aren't exactly desired by the end users. But these situations tend to 
sort themselves out over time.


The SAME thing happens with invaluement's ivmURI domain blacklist. 
OFTEN, a normally legit web site has a CURRENT... LIVE spam infestation, 
where spammers broke into that site and placed spammy content there. 
This has become epidemic. Sure, it is frustrating for everyone, when 
such a site that is being used to send phishing and porn spams... causes 
some of that site's legitimate correspondence to get blocked... but this 
a necessary lesser of evils. The best part is that such a blacklisting 
motivates the site owner to fix their site FASTER. In such a situation, 
the blacklist provided the world a good service, and the resulting 
collateral damage was well justified. The site owner should be 
considered at fault for the collateral damage, not the DNSBL.


I hope this provides some clarity.

--
Rob McEwen
+1 478-475-9032



Re: Uptick in spam

2015-03-30 Thread Rob McEwen

On 3/30/2015 1:19 PM, Kris Deugau wrote:

The cases I
can recall are more along the lines of grey-hat ESPs who pick up a
spammer client for a while,


Kris,

The next time you run across this and think it might be causing a little 
too much collateral damage (in spite of the spamming), let me know 
(off-list) and I'll research it. I can then make adjustments 
accordingly. I'm very responsive to customer feedback.


Thanks!

--
Rob McEwen
+1 478-475-9032



Re: Uptick in spam

2015-03-27 Thread Rob McEwen

On 3/27/2015 10:13 PM, David Jones wrote:

The invaluement RBL is not expensive either and it is awesome.  We pay 
thousands per year for
a Spamhaus feed because of our volume and mailboxes.  The invaluement RBL is 
only hundreds
per year and it's almost as good as Spamhaus Zen.  I have Spamhaus in front of 
invaluement  in
my postfix configuration but I may try flipping the order just to see if it 
will start blocking more
than Spamhaus.


Just to clarify, the two invaluement sender's IP blacklists, ivmSIP and 
ivmSIP/24, --combined-- is not (and will probably not ever be) an 
adequate replacement for Spamhaus's Zen list. So please everyone, don't 
get the idea that you can turn off Zen, add invaluement, and everything 
will be ok. David Jones was NOT saying that... but i just want to make 
sure that nobody mistakenly goes too far with this, beyond what David 
intended.


Having said that... thanks, David, (and others) for your mentioning 
about your success with ivmSIP and ivmSIP/24, where they are helping you 
block much of the spam that slips past Spamhaus, etc.


--
Rob McEwen
 



Re: Ready to throw in the towel on email providing...

2014-07-28 Thread Rob McEwen
On 7/28/2014 12:10 PM, Ted Mittelstaedt wrote:
 Small company with about 6 mailboxes who some consultant gave them a
 song and dance about how Gmail's such a
 better mail service since they don't get any spam 

Ted,

fwiw, I had a situation last year where a friend (not one of my own
clients) called me up asking me about a situation where legit hand-typed
messages to their paid business-class gmail service... were getting blocked.

So I ask, why are you asking me, ask gmail!

He responded by saying that they are a paid business class gmail user,
sending from their own domain name... and when he called his paid
support line at google (again, this isn't the regular free gmail)... the
tech support consultant was not able to lift a finger to help them. No
research.. no SMTP logs... nothing. The paid subscriber was told that
there must be some kind of problem... you'll have to wait this out

In contrast, if one of my mail hosting clients reports that a hand-typed
message to them is blocked, i get the details about the message and
search through the SMTP logs, and report back to them exactly what
happened and fix it if it were something controllable on my end.
(usually there is a more innocent explanation, like the sender making a
typo in the e-mail address, etc)... but I first ASSUME it is my
problem... THEN research it.. then give the client ANSWERS and SOLUTIONS.

PLEASE correct me if I'm wrong as my example above is anecdotal.. but
from what I understand, Google doesn't provide ANSWERS and SOLUTIONS for
situations like this. You just get excuses and delays.

Maybe Google has improved since then?... or maybe my report is not
accurate (but it came to me first hand from a trustworthy source).
Definately double check this. If you can verify that this is true (and
continues to be true)... then use this info as a rebuttal the next time
you have a client talk about leaving you for gmail.

-- 
Rob McEwen
+1 (478) 475-9032



Re: Domain ages (was Re: SPAM from a registrar)

2014-06-10 Thread Rob McEwen
On 6/10/2014 10:21 AM, Axb wrote:
 All  URI BLs I know of (SURBL/URIBL/DBL/Invaluement/etc) check  track
 domain reputation otherwise they'd be unusable.
 Their listings are not blind - they all have their secret sauce to
 process before listing a domain. 

Absolutely. As Axb and KAM and others stated, a very young domain age is
too dangerous to outright block or score high on... but might be a good
factor or good for combining with other rules.

Also, if anyone does see spam that contain domains in the clickable
links where that spam should have been blocked, but was not... then
check the domain contained within the spam again the lookup found at
http://multirbl.valli.org and/or http://mxtoolbox.com/blacklists.aspx
(some months ago, MX Toolbox upgraded their system to check domains
against URI/domain blacklists. In some cases, this could have been a
game of inches where your user caught the tip of the spear and
received the very first spams in a spam campaign that otherwise was
quickly listed by the well known URI BLs. However, you may find that one
or two good URI BLs are simply not implemented in your system--where
that would have made all the difference! Those lookup forms will point
you in the right direction.

The same can also be true for checking sending IPs--then reviewing your
current mix of sender's IP dnsbls (aka RBLs).

Of course, don't fall into the trap of using a BL that catches much, but
has too many FPs. But the list of URI BLs that Axb gave above are all
extremely low-FP URI blacklists.

-- 
Rob McEwen
+1 (478) 475-9032



Re: Domain ages (was Re: SPAM from a registrar)

2014-06-10 Thread Rob McEwen
On 6/10/2014 10:34 AM, Patrick Domack wrote:
 So, we are unwilling to look into any new ideas cause there might be
 an issue? that we haven't scoped or checked into? 

Patrick,

I don't think Axe was arguing against this idea.. I think he was arguing
against irrational exuberance by some who may be taking this idea to the
point of outright blocking (or high scoring) on it, which would generate
significant FPs. His examples were solid real world examples that DO
happen and WOULD FP if this idea were taken too far. But using
extremely-young-domain-age for low scoring or in combination with other
rules could be very helpful.

-- 
Rob McEwen
+1 (478) 475-9032



Re: Domain ages (was Re: SPAM from a registrar)

2014-06-09 Thread Rob McEwen
Domain age is a good metric to factor in. But I'm always fascinated with
some people's desire to block all messages with extremely new domains. 
(NOT saying that this applies to everyone who posted on this thread!)

Keep in mind that many large and famous businesses... who have fairly
good mail sending practices... sometimes launch a new products complete
with links to very newly registered domains. Same is often true for
advertisments for things like rock concerts, etc. Or web sites that deal
with specific events or hot-topic political issues that appeared out of
nowhere. Yes, some of these are UBE. But many are NOT!

These example provide one of the largest source of FPs for all the major
domain/URI blacklists. But the better domain/URI blacklists have good
mechanisms in place to (a) PREVENT... MANY of these from ever becoming
FPs in the first place, and (b) and where those mechanism failed, they
have good triggers/feedback to remove  whitelist such FPs VERY QUICKLY
if/when they do occur.

In contrast, many who might go overboard by outright blocking on
newness... and/or scoring too agressively on newness... may find
too-high FP problems kicking their butts in the long run. And when such
a FP starts happening, they may not have the proper telemetry to
catch/fix it until AFTER much FP damage has happened.

Personally, I think that the real problem here is that some of the most
famous URI/domain blacklists are NOT catching everything and/or NOT
catching everything fast enough... combined with many sys admins failing
to make use of ALL the good and low-FP URI/domain blacklists... where
they 'd see MUCH better results if they were using ALL of the good URI
blacklists! ...but I'm a little biased on this point! :)

-- 
Rob McEwen
+1 (478) 475-9032



Re: Who wants to trade data?

2014-02-06 Thread Rob McEwen
On 2/6/2014 6:59 PM, Noel Butler wrote:
 spams an anti-spam list

so sharing/discussing data/intel about spammers on an anti-spam list...
is spamming? Really?

-- 
Rob McEwen
invaluement.com 



what is that number at the beginning of .cf files signify?

2013-11-14 Thread Rob McEwen
what is that number at the beginning of .cf files signify?

Does that impact SA's actual operation?

Or is that just for human organization of files (how they sort when
browsing them)?

When adding a custom-written .cf file that is made available to the
public, should some kind of naming convention be followed, even if just
for etiquette?

-- 
Rob McEwen
http://dnsbl.invaluement.com/
r...@invaluement.com
+1 (478) 475-9032



Re: Uptick in false negatives - filter check?

2013-11-08 Thread Rob McEwen
On 11/7/2013 6:00 PM, Owen Mehegan wrote:
 Thanks in advance for any advice anyone can offer!

fwiw, of the 4 spam examples, ivmURI had blacklisted one or more domains
in ALL 4 out of 4 samples at least several minutes BEFORE those spams
hit your server (some  days or weeks before).

In a large portion of those (1/2 or more), I'm fairly sure that ivmURI
was the ONLY URI/domain blacklist to have the domain blacklisted at the
time the message hit your network. (I'm unable to verify if DBL had
caught it at that time and/or some of those could have been a game of
inches where ivmURI and other lists had just listed it moments before
and it would be somewhat of a propagation issue... but, overall, I think
if I provided the date/times that these were blacklisted on ivmURI...
that assertion would check out and the raw data would be rather
impressive!)

If you keep seeing these, check the domains on multirbl.valli.org ...and
you'll see in real time what I'm talking about!

-- 
Rob McEwen
http://dnsbl.invaluement.com/
r...@invaluement.com
+1 (478) 475-9032



Re: KAM pccc URIBL questions

2013-10-07 Thread Rob McEwen
On 10/7/2013 7:42 PM, Raymond Dijkxhoorn wrote:
 This is harming more then it does good. But its your list so your
 rules ;) I would not want to use it to filter my mails with it but hey

Since this is in its early development, it is probably too early to
judge it too much. But from what I've read in this discussion, it is
light years away from the current major URI/domain blacklists out
there (SURBL, URIBL, ivmURI, DBL)... BUT... Kevin  is  brilliant so who
knows what it might eventually become?

ALSO...There is an argument that a more-aggressive-than-normal AND
low-scoring URI list may be helpful? In that sense, URIBL.com has
traditionally been considered slightly more aggressive than the other
lists mentioned above... SLIGHTLY! Maybe something much MORE aggressive,
intended for very low scoring... would be useful? (this would be
situations where bayes or checksum content filters add points to the
spam score combined with such an aggressive URI list putting the message
over the top... but then skipping blocking a legit message with this
URI because it didn't have the other content points added and thus
didn't score high enough--at least that is the idea)

But I can't help but think that SOME reading this thread haven't even
tried/implemented even all the zero-cost options for the (already
matured) lists I mentioned (where applicable)?

-- 
Rob McEwen
http://dnsbl.invaluement.com/
r...@invaluement.com
+1 (478) 475-9032



  1   2   3   4   >