Hello Everyone!

First, my apologies for not doing my Lurk Time here - I only started
subscribing to this list today.

Unfortunately, I do not have any lists to share as I have none, but I do
have some ideas...

We all know there have long been RBL's for spam sources on the net and as an
SMTP admin for a major ISP, they are invaluable at keeping the spam out by
preventing tcp/25 connections from blacklisted IP addresses.

Well, I am also a DNS admin, and in the past, I have had to block queries to
certain domains (mostly dealing with child porno) by court order from
various states in the U.S.  

I have one suggestion on format of the data which is shared.  It should
include the domain, the host, and the actual URL at a minimum, in CSV
format.  For example:

Badhost.com, www.badhost.com, http://www.badhost.com/badfile.exe

Then, with Rsync, this data could be shared in near-real time and handled by
administrators to suite their needs.  By parsing the data, A DNS
administrator could create scripts and choose to set badhost.com to
something like this in their named.conf:

zone "badhost.com" {
        type master;
        file "empty.zone";
        allow-update { none; };
};

And empty.zone would look something like this:

@ 10800 IN SOA ns1.mydomain.com. root.mydomain.com. (
               1 3600 1200 604800 10800 )
@ 10800 IN NS ns1.mydomain.com.

The end result is that a DNS query for any host in the list would return
nothing.

If the data were vast enough, and if enough administrators subscribed and
used this 'blacklist', it would have a real effect against end users hitting
bad URL's.  I have no problems blocking bad domains, just as maintainers of
spam RBL's have no problems blocking bad IP's.  It is up to the ISP of the
bad host to take care of the problem to become unblocked.

As far as http or hxxp, it doesn't matter to me, I would only handle this
information on the backend.




-----Original Message-----
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Chris Lee
Sent: Thursday, August 28, 2008 5:54 PM
To: botnets@whitestar.linuxbox.org
Subject: Re: [botnets] URL formats

hxxp seems to be advantageous for a few reasons:
  1. you can still cut and paste the url
  2. the protocol handlers won't load it up if you accidently click  
on it
  3. you can add a protocol handler for hxxp for whatever you want
  4. easier to recognize domains and patterns (rather than rotted urls)
  5. already widely accepted in spam fighting groups
  6. trivial to do and undo with no exception cases

I figured I'd put down my thoughts to try to help a standard to move  
forward.


On Aug 28, 2008, at 7:07 PM, silky wrote:

> On Fri, Aug 29, 2008 at 3:32 AM, Chris Burton <[EMAIL PROTECTED]> wrote:
>> Hi,
>> I was wondering if it would be more helpful if we could propose a  
>> "standard"
>> for posting broken URLs with some form of start/end indicator to  
>> allow
>> easier automated processing from the listings?
>
> I was thinking that it would be nice to post them just rot13'd. Still
> trivially decoded (i use leetkey add-in in ff) but not picked up by
> indexers/etc. Advantage is that it can still be searched for common
> patterns.
>
>
>> ChrisB.
>
>
> -- 
> noon silky
> http://www.themonkeynet.com/armada/
> _______________________________________________
> botnets@, the public's dumping ground for maliciousness
> All list and server information are public and available to law  
> enforcement upon request.
> http://www.whitestar.linuxbox.org/mailman/listinfo/botnets

_______________________________________________
botnets@, the public's dumping ground for maliciousness
All list and server information are public and available to law enforcement
upon request.
http://www.whitestar.linuxbox.org/mailman/listinfo/botnets

_______________________________________________
botnets@, the public's dumping ground for maliciousness
All list and server information are public and available to law enforcement 
upon request.
http://www.whitestar.linuxbox.org/mailman/listinfo/botnets

Reply via email to