RE: Monitoring other people's sites (Was: Website for ipv6.level3.com returns HTTP/1.1 500 Internal Server Error)

2012-03-22 Thread Frank Bulk
I and my customers users IPv6-enabled sites.  If it doesn't work (a hopefully 
their web browser uses HE) I want to know, and know when it happens.  Yes, many 
sites aren't monitoring their own IPv6-connected content, but I've had 
reasonably good success privately letting them know when it's down.  And 
communicating to them when it's down lets them know that people care and want 
to access their IPv6-enabled content.  Last, monitoring IPv6 access to many 
different sites brings our own connectivity issues to the surface as they arise 
-- we had one inside Level3's network last week Friday and it was resolved 
about 18 hours later.  If we had not monitored it's possible it would be much 
longer before it was discovered and troubleshot through the regular sequence of 
events.

Frank

-Original Message-
From: Jeroen Massar [mailto:jer...@unfix.org] 
Sent: Tuesday, March 20, 2012 9:54 AM
To: vinny_abe...@dell.com
Cc: nanog@nanog.org
Subject: Monitoring other people's sites (Was: Website for ipv6.level3.com 
returns HTTP/1.1 500 Internal Server Error)

snip

 And for the few folks putting nagios's on other people's sites, they
obviously do not understand that even if the alarm goes off that
something is broken that they cannot fix it anyway, thus why bother...






Re: Monitoring other people's sites (Was: Website for ipv6.level3.com returns HTTP/1.1 500 Internal Server Error)

2012-03-21 Thread Jeroen Massar
On 2012-03-20 16:53 , Nick Hilliard wrote:
 On 20/03/2012 14:54, Jeroen Massar wrote:
 For everybody who is monitoring other people's websites, please please
 please, monitor something static like /robots.txt as that can be
 statically served and is kinda appropriate as it is intended for robots.
 
 Depends on what you are monitoring.  If you're looking for layer 4 ipv6
 connectivity then robots.txt is fine.  If you're trying to determine
 whether a site is serving active content on ipv6 and not serving http
 errors, then it's pretty pointless to monitor robots.txt - you need to
 monitor /.

And as can be seen with the monitoring of ipv6.level3.com it will tell
you 'it is broken' but as the person who is monitoring has no relation
or contact with them, it only leads to public complaints which do not
get resolved

If the site themselves cannot be arsed to monitor their own, then why
would you bother to do so.

Indeed, I agree that it can be useful, especially as an access ISP, to
monitor popular websites so that you know that you can reach them, but
that does not mean you need to pull large amounts of data.
(for determining MTU issues yes, but likely you have a full 1500 path
anyway thus these should as good as possible not happen anyway)

But unless you have a contact at the site it will be tough to resolve
the issue anyway.

 Oh and of course do set the User-Agent to something logical and to be
 super nice include a contact address so that people who do check their
 logs once in a while for fishy things they at least know what is
 happening there and that it is not a process run afoul or something.
 
 Good policy, yes.  Some robots do this but others don't.
 
 Of course, asking before doing tends to be a good idea too.
 
 Depends on the scale.  I'm not going to ask permission to poll someone
 else's site every 5 minutes, and I would be surprised if they asked me the
 same.  OTOH, if they were polling to the point that it was causing issues,
 that might be different.

I was not talking about that low rate, not a lot of people will notice
that, but the 1000qps from 500 sources was quite noticed and thus at
first they got blocked, then we tried to find out who was doing it, and
then they repointed to robots.txt, unblocked them and all was fine.

 The IPv6 Internet already consists way too much out of monitoring by
 pulling pages and doing pings...
 
 way too much for what?  IPv6 is not widely adopted.

In comparison to real traffic. There has been a saying since the 6bone
days already that IPv6 is just ICMPv6...

 Fortunately that should heavily change in a few months.
 
 We've been saying this for years.  World IPv6 day 2012 will come and go,
 and things are unlikely to change a whole lot.  The only thing that World
 IPv6 day 2012 will ensure is that people whose ipv6 configuration actively
 interferes with their daily Internet usage will be self-flagged and their
 configuration issues can be dealt with.

Fully agree, but at least at that point nobody will be able to claim
that they can't deploy IPv6 on the access side as there is no content ;)

  (who noticed a certain sh company performing latency checks against
 one of his sites, which was no problem, but the fact that they where
 causing almost more hits/traffic/load than normal clients was a bit on
 the much side
 
 If that web page is configured to be as top-heavy as this, then I'd suggest
 putting a cache in front of it. nginx is good for this sort of thing.

nginx does not help if your content is not cacheable by nginx, for
instance if you simply show the IP address of the client and if they
thus have IPv6 or IPv4.

In our case, indeed, everything that is static is served by nginx, which
is why hammering on /robots.txt is not an issue at all...

On 2012-03-20 21:45 , Charles N Wyble wrote:
 On 03/20/2012 09:54 AM, Jeroen Massar wrote:
 On 2012-03-20 15:40 , vinny_abe...@dell.com wrote:
 
 For everybody who is monitoring other people's websites, please
 please please, monitor something static like /robots.txt as that
 can be statically served and is kinda appropriate as it is intended
 for robots.
 
 This could provide a false positive if one is interested in ensuring 
 that the full application stack is working.

As stated above and given the example of the original subject of
ipv6.level3.com, what exactly are you going to do when it does not?

And again, if the owner does not care, why should you?

Also, maybe they do a redesign of the site and remove the keywords or
other metrics you are looking for. It is not your problem to monitor it
for them, unless they hire you to do so of course.

 Oh and of course do set the User-Agent to something logical and to
 be super nice include a contact address so that people who do check
 their logs once in a while for fishy things they at least know what
 is happening there and that it is not a process run afoul or
 something.
 
 A server side process? Or client side?

Take a guess what something that 

Re: Monitoring other people's sites (Was: Website for ipv6.level3.com returns HTTP/1.1 500 Internal Server Error)

2012-03-20 Thread Nick Hilliard
On 20/03/2012 14:54, Jeroen Massar wrote:
 For everybody who is monitoring other people's websites, please please
 please, monitor something static like /robots.txt as that can be
 statically served and is kinda appropriate as it is intended for robots.

Depends on what you are monitoring.  If you're looking for layer 4 ipv6
connectivity then robots.txt is fine.  If you're trying to determine
whether a site is serving active content on ipv6 and not serving http
errors, then it's pretty pointless to monitor robots.txt - you need to
monitor /.

 Oh and of course do set the User-Agent to something logical and to be
 super nice include a contact address so that people who do check their
 logs once in a while for fishy things they at least know what is
 happening there and that it is not a process run afoul or something.

Good policy, yes.  Some robots do this but others don't.

 Of course, asking before doing tends to be a good idea too.

Depends on the scale.  I'm not going to ask permission to poll someone
else's site every 5 minutes, and I would be surprised if they asked me the
same.  OTOH, if they were polling to the point that it was causing issues,
that might be different.

 The IPv6 Internet already consists way too much out of monitoring by
 pulling pages and doing pings...

way too much for what?  IPv6 is not widely adopted.

 Fortunately that should heavily change in a few months.

We've been saying this for years.  World IPv6 day 2012 will come and go,
and things are unlikely to change a whole lot.  The only thing that World
IPv6 day 2012 will ensure is that people whose ipv6 configuration actively
interferes with their daily Internet usage will be self-flagged and their
configuration issues can be dealt with.

  (who noticed a certain sh company performing latency checks against
 one of his sites, which was no problem, but the fact that they where
 causing almost more hits/traffic/load than normal clients was a bit on
 the much side

If that web page is configured to be as top-heavy as this, then I'd suggest
putting a cache in front of it. nginx is good for this sort of thing.

Nick



Re: Monitoring other people's sites (Was: Website for ipv6.level3.com returns HTTP/1.1 500 Internal Server Error)

2012-03-20 Thread Jason Hellenthal


On Tue, Mar 20, 2012 at 03:54:13PM +0100, Jeroen Massar wrote:
 On 2012-03-20 15:40 , vinny_abe...@dell.com wrote:
  FYI - it's also the main IPv4 site, not just IPv6... although I'm
  unsure if it's the same issue.
  
  I was monitoring availability as a point of reference for my network
  and started receiving 500 errors recently as well that tripped up the
  monitoring system, even though the page comes up in any browser I
  try.
  
  GET / HTTP/1.1 User-Agent: Mozilla/4.0 (compatible; MSIE 4.01;
  Windows NT)
 
 For everybody who is monitoring other people's websites, please please
 please, monitor something static like /robots.txt as that can be
 statically served and is kinda appropriate as it is intended for robots.
 Oh and of course do set the User-Agent to something logical and to be
 super nice include a contact address so that people who do check their
 logs once in a while for fishy things they at least know what is
 happening there and that it is not a process run afoul or something.
 
 Of course, asking before doing tends to be a good idea too.
 
 The IPv6 Internet already consists way too much out of monitoring by
 pulling pages and doing pings...
 
 Fortunately that should heavily change in a few months.
 
 Greets,
  Jeroen
 
  (who noticed a certain sh company performing latency checks against
 one of his sites, which was no problem, but the fact that they where
 causing almost more hits/traffic/load than normal clients was a bit on
 the much side, them pulling robots.txt solved their problem to be able
 to check if their IPv6 worked fine and the load issue on the server side
 was gone too as nginx happily serves little robots.txt's at great speed
 from cache ;)
 
  And for the few folks putting nagios's on other people's sites, they
 obviously do not understand that even if the alarm goes off that
 something is broken that they cannot fix it anyway, thus why bother...

I agree! leave the monitoring for those that are hired to do so. Using
someone elses server to verify that your ipv6 connectivity works should
just strictly get your traffic dropped or null-routed with an alert sent
to your provider.

ping6 your provider... wget -6 your provider but beyond that you, most
likely cannot fix it...

-- 
;s =;



Re: Monitoring other people's sites (Was: Website for ipv6.level3.com returns HTTP/1.1 500 Internal Server Error)

2012-03-20 Thread Charles N Wyble
On 03/20/2012 09:54 AM, Jeroen Massar wrote:
 On 2012-03-20 15:40 , vinny_abe...@dell.com wrote:

 For everybody who is monitoring other people's websites, please please
 please, monitor something static like /robots.txt as that can be
 statically served and is kinda appropriate as it is intended for robots.

This could provide a false positive if one is interested in ensuring
that the full application stack is working.

 Oh and of course do set the User-Agent to something logical and to be
 super nice include a contact address so that people who do check their
 logs once in a while for fishy things they at least know what is
 happening there and that it is not a process run afoul or something.

A server side process? Or client side? If the client side monitoring is
too aggressive , then your rate limiting firewall rules should kick in
and block it. If you don't have a rate limiting firewall on your web
server, (on the server itself, not in front of it) then you have bigger
problems.

 Of course, asking before doing tends to be a good idea too.


If you are running a public service, expect it to get
monitored/attacked/probed etc. If you don't want traffic from certain
sources then block it.

 The IPv6 Internet already consists way too much out of monitoring by
 pulling pages and doing pings...

Who made you the arbiter of acceptable automated traffic levels?



  (who noticed a certain sh company performing latency checks against
 one of his sites, which was no problem, but the fact that they where
 causing almost more hits/traffic/load than normal clients was a bit on
 the much side,

Again. Use a firewall and limit them if the traffic isn't in line with
your site policies.

  And for the few folks putting nagios's on other people's sites, they
 obviously do not understand that even if the alarm goes off that
 something is broken that they cannot fix it anyway, thus why bother...

You obviously do not understand why people are implementing these
monitors. It's to serve as a canary for v6 connectivity issues. If I was
implementing a monitor like this, I'd use the following logic:

HTTP 200 returned via v4/v6 == all is well
HTTP 200 returned via v4 or v6 , no HTTP code returned via v4 or v6 (ie
one path works) ==  v6/v4 potentially broken.
no HTTP code returned via either method == end site problem. nothing we
can do. don't alert.

Presumably you'd also implement a TCP 80 check as well.